Dec 05 10:35:18.411084 master-0 systemd[1]: Starting Kubernetes Kubelet... Dec 05 10:35:18.557297 master-0 kubenswrapper[4752]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:35:18.557297 master-0 kubenswrapper[4752]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 05 10:35:18.557297 master-0 kubenswrapper[4752]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:35:18.557297 master-0 kubenswrapper[4752]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:35:18.557297 master-0 kubenswrapper[4752]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 05 10:35:18.557297 master-0 kubenswrapper[4752]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:35:18.558351 master-0 kubenswrapper[4752]: I1205 10:35:18.557791 4752 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 05 10:35:18.560073 master-0 kubenswrapper[4752]: W1205 10:35:18.560046 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:35:18.560073 master-0 kubenswrapper[4752]: W1205 10:35:18.560062 4752 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:35:18.560073 master-0 kubenswrapper[4752]: W1205 10:35:18.560067 4752 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:35:18.560073 master-0 kubenswrapper[4752]: W1205 10:35:18.560072 4752 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:35:18.560073 master-0 kubenswrapper[4752]: W1205 10:35:18.560078 4752 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560083 4752 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560088 4752 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560092 4752 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560096 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560100 4752 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560105 4752 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560108 4752 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560112 4752 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560116 4752 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560120 4752 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560124 4752 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560128 4752 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560131 4752 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560135 4752 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560138 4752 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560142 4752 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560146 4752 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560149 4752 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:35:18.560235 master-0 kubenswrapper[4752]: W1205 10:35:18.560153 4752 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560158 4752 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560163 4752 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560166 4752 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560170 4752 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560174 4752 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560178 4752 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560182 4752 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560185 4752 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560189 4752 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560193 4752 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560197 4752 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560202 4752 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560206 4752 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560209 4752 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560213 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560217 4752 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560221 4752 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560224 4752 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:35:18.560792 master-0 kubenswrapper[4752]: W1205 10:35:18.560228 4752 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560231 4752 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560235 4752 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560239 4752 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560242 4752 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560246 4752 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560250 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560254 4752 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560258 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560261 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560265 4752 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560269 4752 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560272 4752 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560276 4752 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560279 4752 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560283 4752 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560286 4752 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560290 4752 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560294 4752 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560297 4752 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:35:18.561311 master-0 kubenswrapper[4752]: W1205 10:35:18.560301 4752 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: W1205 10:35:18.560305 4752 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: W1205 10:35:18.560308 4752 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: W1205 10:35:18.560312 4752 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: W1205 10:35:18.560316 4752 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: W1205 10:35:18.560320 4752 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: W1205 10:35:18.560324 4752 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: W1205 10:35:18.560329 4752 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: W1205 10:35:18.560334 4752 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: W1205 10:35:18.560338 4752 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560594 4752 flags.go:64] FLAG: --address="0.0.0.0" Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560605 4752 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560613 4752 flags.go:64] FLAG: --anonymous-auth="true" Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560618 4752 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560624 4752 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560628 4752 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560634 4752 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560639 4752 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560644 4752 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560648 4752 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560653 4752 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 05 10:35:18.561915 master-0 kubenswrapper[4752]: I1205 10:35:18.560657 4752 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560661 4752 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560665 4752 flags.go:64] FLAG: --cgroup-root="" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560669 4752 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560674 4752 flags.go:64] FLAG: --client-ca-file="" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560678 4752 flags.go:64] FLAG: --cloud-config="" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560682 4752 flags.go:64] FLAG: --cloud-provider="" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560686 4752 flags.go:64] FLAG: --cluster-dns="[]" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560691 4752 flags.go:64] FLAG: --cluster-domain="" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560695 4752 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560699 4752 flags.go:64] FLAG: --config-dir="" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560703 4752 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560708 4752 flags.go:64] FLAG: --container-log-max-files="5" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560713 4752 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560717 4752 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560722 4752 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560727 4752 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560731 4752 flags.go:64] FLAG: --contention-profiling="false" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560735 4752 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560744 4752 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560749 4752 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560753 4752 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560759 4752 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560764 4752 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560769 4752 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 05 10:35:18.562500 master-0 kubenswrapper[4752]: I1205 10:35:18.560774 4752 flags.go:64] FLAG: --enable-load-reader="false" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560778 4752 flags.go:64] FLAG: --enable-server="true" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560783 4752 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560790 4752 flags.go:64] FLAG: --event-burst="100" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560794 4752 flags.go:64] FLAG: --event-qps="50" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560799 4752 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560803 4752 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560807 4752 flags.go:64] FLAG: --eviction-hard="" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560815 4752 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560819 4752 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560823 4752 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560828 4752 flags.go:64] FLAG: --eviction-soft="" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560832 4752 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560836 4752 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560840 4752 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560844 4752 flags.go:64] FLAG: --experimental-mounter-path="" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560849 4752 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560853 4752 flags.go:64] FLAG: --fail-swap-on="true" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560857 4752 flags.go:64] FLAG: --feature-gates="" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560862 4752 flags.go:64] FLAG: --file-check-frequency="20s" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560867 4752 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560871 4752 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560876 4752 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560881 4752 flags.go:64] FLAG: --healthz-port="10248" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560885 4752 flags.go:64] FLAG: --help="false" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560889 4752 flags.go:64] FLAG: --hostname-override="" Dec 05 10:35:18.563183 master-0 kubenswrapper[4752]: I1205 10:35:18.560895 4752 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560900 4752 flags.go:64] FLAG: --http-check-frequency="20s" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560904 4752 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560908 4752 flags.go:64] FLAG: --image-credential-provider-config="" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560912 4752 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560916 4752 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560923 4752 flags.go:64] FLAG: --image-service-endpoint="" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560928 4752 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560932 4752 flags.go:64] FLAG: --kube-api-burst="100" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560936 4752 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560940 4752 flags.go:64] FLAG: --kube-api-qps="50" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560945 4752 flags.go:64] FLAG: --kube-reserved="" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560949 4752 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560953 4752 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560957 4752 flags.go:64] FLAG: --kubelet-cgroups="" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560961 4752 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560965 4752 flags.go:64] FLAG: --lock-file="" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560969 4752 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560973 4752 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560977 4752 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560984 4752 flags.go:64] FLAG: --log-json-split-stream="false" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560988 4752 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560992 4752 flags.go:64] FLAG: --log-text-split-stream="false" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.560996 4752 flags.go:64] FLAG: --logging-format="text" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.561000 4752 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 05 10:35:18.563985 master-0 kubenswrapper[4752]: I1205 10:35:18.561004 4752 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561009 4752 flags.go:64] FLAG: --manifest-url="" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561013 4752 flags.go:64] FLAG: --manifest-url-header="" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561018 4752 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561023 4752 flags.go:64] FLAG: --max-open-files="1000000" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561028 4752 flags.go:64] FLAG: --max-pods="110" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561032 4752 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561038 4752 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561042 4752 flags.go:64] FLAG: --memory-manager-policy="None" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561046 4752 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561051 4752 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561055 4752 flags.go:64] FLAG: --node-ip="192.168.32.10" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561059 4752 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561068 4752 flags.go:64] FLAG: --node-status-max-images="50" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561073 4752 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561077 4752 flags.go:64] FLAG: --oom-score-adj="-999" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561081 4752 flags.go:64] FLAG: --pod-cidr="" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561086 4752 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a70b2a95140d1e90978f36cc9889013ae34bd232662c5424002274385669ed9" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561092 4752 flags.go:64] FLAG: --pod-manifest-path="" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561096 4752 flags.go:64] FLAG: --pod-max-pids="-1" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561100 4752 flags.go:64] FLAG: --pods-per-core="0" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561105 4752 flags.go:64] FLAG: --port="10250" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561110 4752 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561115 4752 flags.go:64] FLAG: --provider-id="" Dec 05 10:35:18.564889 master-0 kubenswrapper[4752]: I1205 10:35:18.561119 4752 flags.go:64] FLAG: --qos-reserved="" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561123 4752 flags.go:64] FLAG: --read-only-port="10255" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561127 4752 flags.go:64] FLAG: --register-node="true" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561131 4752 flags.go:64] FLAG: --register-schedulable="true" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561135 4752 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561142 4752 flags.go:64] FLAG: --registry-burst="10" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561146 4752 flags.go:64] FLAG: --registry-qps="5" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561150 4752 flags.go:64] FLAG: --reserved-cpus="" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561155 4752 flags.go:64] FLAG: --reserved-memory="" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561160 4752 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561164 4752 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561168 4752 flags.go:64] FLAG: --rotate-certificates="false" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561172 4752 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561177 4752 flags.go:64] FLAG: --runonce="false" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561181 4752 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561185 4752 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561191 4752 flags.go:64] FLAG: --seccomp-default="false" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561195 4752 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561199 4752 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561204 4752 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561208 4752 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561212 4752 flags.go:64] FLAG: --storage-driver-password="root" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561216 4752 flags.go:64] FLAG: --storage-driver-secure="false" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561220 4752 flags.go:64] FLAG: --storage-driver-table="stats" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561224 4752 flags.go:64] FLAG: --storage-driver-user="root" Dec 05 10:35:18.565579 master-0 kubenswrapper[4752]: I1205 10:35:18.561228 4752 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561233 4752 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561237 4752 flags.go:64] FLAG: --system-cgroups="" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561241 4752 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561248 4752 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561252 4752 flags.go:64] FLAG: --tls-cert-file="" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561256 4752 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561261 4752 flags.go:64] FLAG: --tls-min-version="" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561265 4752 flags.go:64] FLAG: --tls-private-key-file="" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561269 4752 flags.go:64] FLAG: --topology-manager-policy="none" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561273 4752 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561278 4752 flags.go:64] FLAG: --topology-manager-scope="container" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561282 4752 flags.go:64] FLAG: --v="2" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561287 4752 flags.go:64] FLAG: --version="false" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561293 4752 flags.go:64] FLAG: --vmodule="" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561298 4752 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: I1205 10:35:18.561302 4752 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: W1205 10:35:18.561394 4752 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: W1205 10:35:18.561400 4752 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: W1205 10:35:18.561405 4752 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: W1205 10:35:18.561410 4752 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: W1205 10:35:18.561414 4752 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:35:18.566274 master-0 kubenswrapper[4752]: W1205 10:35:18.561432 4752 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561438 4752 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561441 4752 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561445 4752 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561449 4752 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561452 4752 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561456 4752 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561460 4752 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561463 4752 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561467 4752 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561471 4752 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561475 4752 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561478 4752 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561482 4752 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561486 4752 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561490 4752 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561494 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561497 4752 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561502 4752 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561506 4752 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:35:18.566917 master-0 kubenswrapper[4752]: W1205 10:35:18.561510 4752 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561513 4752 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561517 4752 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561521 4752 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561524 4752 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561528 4752 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561531 4752 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561535 4752 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561539 4752 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561542 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561546 4752 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561549 4752 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561553 4752 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561559 4752 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561564 4752 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561569 4752 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561573 4752 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561577 4752 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561581 4752 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:35:18.567504 master-0 kubenswrapper[4752]: W1205 10:35:18.561585 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561588 4752 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561592 4752 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561595 4752 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561599 4752 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561603 4752 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561607 4752 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561611 4752 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561614 4752 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561618 4752 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561621 4752 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561625 4752 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561629 4752 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561632 4752 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561636 4752 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561640 4752 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561644 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561648 4752 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561651 4752 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561655 4752 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:35:18.568195 master-0 kubenswrapper[4752]: W1205 10:35:18.561658 4752 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:35:18.568782 master-0 kubenswrapper[4752]: W1205 10:35:18.561665 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:35:18.568782 master-0 kubenswrapper[4752]: W1205 10:35:18.561669 4752 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:35:18.568782 master-0 kubenswrapper[4752]: W1205 10:35:18.561672 4752 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:35:18.568782 master-0 kubenswrapper[4752]: W1205 10:35:18.561676 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:35:18.568782 master-0 kubenswrapper[4752]: W1205 10:35:18.561679 4752 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:35:18.568782 master-0 kubenswrapper[4752]: W1205 10:35:18.561684 4752 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:35:18.568782 master-0 kubenswrapper[4752]: W1205 10:35:18.561688 4752 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:35:18.568782 master-0 kubenswrapper[4752]: I1205 10:35:18.561893 4752 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false StreamingCollectionEncodingToJSON:false StreamingCollectionEncodingToProtobuf:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 10:35:18.571336 master-0 kubenswrapper[4752]: I1205 10:35:18.571255 4752 server.go:491] "Kubelet version" kubeletVersion="v1.31.13" Dec 05 10:35:18.571336 master-0 kubenswrapper[4752]: I1205 10:35:18.571300 4752 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571391 4752 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571401 4752 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571407 4752 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571412 4752 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571417 4752 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571441 4752 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571445 4752 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571449 4752 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571453 4752 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571457 4752 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571461 4752 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571465 4752 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571469 4752 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571474 4752 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571478 4752 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571482 4752 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571486 4752 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571491 4752 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571495 4752 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:35:18.571463 master-0 kubenswrapper[4752]: W1205 10:35:18.571499 4752 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571503 4752 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571507 4752 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571512 4752 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571516 4752 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571520 4752 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571524 4752 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571528 4752 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571533 4752 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571539 4752 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571543 4752 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571548 4752 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571552 4752 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571557 4752 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571561 4752 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571565 4752 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571569 4752 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571572 4752 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571576 4752 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:35:18.572087 master-0 kubenswrapper[4752]: W1205 10:35:18.571580 4752 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571584 4752 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571588 4752 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571591 4752 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571595 4752 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571599 4752 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571604 4752 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571609 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571613 4752 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571617 4752 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571621 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571626 4752 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571630 4752 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571633 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571637 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571642 4752 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571646 4752 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571650 4752 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571654 4752 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:35:18.572631 master-0 kubenswrapper[4752]: W1205 10:35:18.571657 4752 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571661 4752 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571664 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571668 4752 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571672 4752 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571675 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571679 4752 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571682 4752 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571686 4752 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571691 4752 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571695 4752 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571698 4752 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571702 4752 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571705 4752 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: W1205 10:35:18.571709 4752 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:35:18.573244 master-0 kubenswrapper[4752]: I1205 10:35:18.571715 4752 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false StreamingCollectionEncodingToJSON:false StreamingCollectionEncodingToProtobuf:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571859 4752 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571868 4752 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571872 4752 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571876 4752 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571880 4752 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571884 4752 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571888 4752 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571892 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571896 4752 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571901 4752 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571906 4752 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571910 4752 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571916 4752 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571920 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571923 4752 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571927 4752 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571931 4752 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571934 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:35:18.573688 master-0 kubenswrapper[4752]: W1205 10:35:18.571938 4752 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571942 4752 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571946 4752 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571949 4752 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571953 4752 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571957 4752 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571961 4752 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571965 4752 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571969 4752 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571972 4752 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571976 4752 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571980 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571984 4752 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571987 4752 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571991 4752 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.571996 4752 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.572001 4752 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.572005 4752 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.572010 4752 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.572014 4752 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:35:18.574287 master-0 kubenswrapper[4752]: W1205 10:35:18.572019 4752 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572023 4752 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572027 4752 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572032 4752 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572036 4752 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572040 4752 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572044 4752 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572048 4752 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572052 4752 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572056 4752 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572060 4752 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572064 4752 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572068 4752 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572072 4752 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572075 4752 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572079 4752 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572083 4752 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572087 4752 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572091 4752 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572095 4752 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:35:18.574921 master-0 kubenswrapper[4752]: W1205 10:35:18.572099 4752 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572103 4752 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572108 4752 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572113 4752 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572139 4752 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572143 4752 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572147 4752 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572151 4752 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572155 4752 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572159 4752 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572163 4752 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572166 4752 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572170 4752 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: W1205 10:35:18.572174 4752 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: I1205 10:35:18.572179 4752 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false StreamingCollectionEncodingToJSON:false StreamingCollectionEncodingToProtobuf:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 10:35:18.575464 master-0 kubenswrapper[4752]: I1205 10:35:18.572611 4752 server.go:940] "Client rotation is on, will bootstrap in background" Dec 05 10:35:18.575884 master-0 kubenswrapper[4752]: I1205 10:35:18.574325 4752 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Dec 05 10:35:18.578216 master-0 kubenswrapper[4752]: I1205 10:35:18.577638 4752 server.go:997] "Starting client certificate rotation" Dec 05 10:35:18.578216 master-0 kubenswrapper[4752]: I1205 10:35:18.577670 4752 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 05 10:35:18.578216 master-0 kubenswrapper[4752]: I1205 10:35:18.577913 4752 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 05 10:35:18.583645 master-0 kubenswrapper[4752]: I1205 10:35:18.583615 4752 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 10:35:18.585829 master-0 kubenswrapper[4752]: I1205 10:35:18.585792 4752 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 10:35:18.586138 master-0 kubenswrapper[4752]: E1205 10:35:18.586076 4752 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.sno.openstack.lab:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:18.594082 master-0 kubenswrapper[4752]: I1205 10:35:18.594042 4752 log.go:25] "Validated CRI v1 runtime API" Dec 05 10:35:18.597173 master-0 kubenswrapper[4752]: I1205 10:35:18.597134 4752 log.go:25] "Validated CRI v1 image API" Dec 05 10:35:18.599245 master-0 kubenswrapper[4752]: I1205 10:35:18.599216 4752 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 05 10:35:18.601818 master-0 kubenswrapper[4752]: I1205 10:35:18.601776 4752 fs.go:135] Filesystem UUIDs: map[354073b4-663b-4c82-b811-53676c745317:/dev/vda3 7B77-95E7:/dev/vda2 910678ff-f77e-4a7d-8d53-86f2ac47a823:/dev/vda4] Dec 05 10:35:18.601818 master-0 kubenswrapper[4752]: I1205 10:35:18.601810 4752 fs.go:136] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0}] Dec 05 10:35:18.623711 master-0 kubenswrapper[4752]: I1205 10:35:18.623353 4752 manager.go:217] Machine: {Timestamp:2025-12-05 10:35:18.621705897 +0000 UTC m=+0.161829246 CPUVendorID:AuthenticAMD NumCores:16 NumPhysicalCores:1 NumSockets:16 CpuFrequency:2799998 MemoryCapacity:50514153472 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:9319397e1208473fb2e5a8b942c60cc5 SystemUUID:9319397e-1208-473f-b2e5-a8b942c60cc5 BootID:759865f8-012e-4147-b61d-8ab8adc2639f Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:10102833152 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:25257078784 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:25257074688 Type:vfs Inodes:6166278 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none} 252:16:{Name:vdb Major:252 Minor:16 Size:21474836480 Scheduler:none} 252:32:{Name:vdc Major:252 Minor:32 Size:21474836480 Scheduler:none} 252:48:{Name:vdd Major:252 Minor:48 Size:21474836480 Scheduler:none} 252:64:{Name:vde Major:252 Minor:64 Size:21474836480 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:9e:81:f6:10 Speed:0 Mtu:9000} {Name:eth0 MacAddress:fa:16:9e:81:f6:10 Speed:-1 Mtu:9000} {Name:eth1 MacAddress:fa:16:3e:5e:6c:eb Speed:-1 Mtu:9000} {Name:eth2 MacAddress:fa:16:3e:5e:ad:70 Speed:-1 Mtu:9000} {Name:ovs-system MacAddress:6e:04:ec:85:c4:7f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:50514153472 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[12] Caches:[{Id:12 Size:32768 Type:Data Level:1} {Id:12 Size:32768 Type:Instruction Level:1} {Id:12 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:12 Size:16777216 Type:Unified Level:3}] SocketID:12 BookID: DrawerID:} {Id:0 Threads:[13] Caches:[{Id:13 Size:32768 Type:Data Level:1} {Id:13 Size:32768 Type:Instruction Level:1} {Id:13 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:13 Size:16777216 Type:Unified Level:3}] SocketID:13 BookID: DrawerID:} {Id:0 Threads:[14] Caches:[{Id:14 Size:32768 Type:Data Level:1} {Id:14 Size:32768 Type:Instruction Level:1} {Id:14 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:14 Size:16777216 Type:Unified Level:3}] SocketID:14 BookID: DrawerID:} {Id:0 Threads:[15] Caches:[{Id:15 Size:32768 Type:Data Level:1} {Id:15 Size:32768 Type:Instruction Level:1} {Id:15 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:15 Size:16777216 Type:Unified Level:3}] SocketID:15 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 05 10:35:18.623711 master-0 kubenswrapper[4752]: I1205 10:35:18.623678 4752 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 05 10:35:18.623857 master-0 kubenswrapper[4752]: I1205 10:35:18.623827 4752 manager.go:233] Version: {KernelVersion:5.14.0-427.100.1.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202511170715-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 05 10:35:18.624445 master-0 kubenswrapper[4752]: I1205 10:35:18.624395 4752 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 05 10:35:18.624653 master-0 kubenswrapper[4752]: I1205 10:35:18.624605 4752 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 05 10:35:18.624869 master-0 kubenswrapper[4752]: I1205 10:35:18.624647 4752 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"master-0","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 05 10:35:18.624930 master-0 kubenswrapper[4752]: I1205 10:35:18.624912 4752 topology_manager.go:138] "Creating topology manager with none policy" Dec 05 10:35:18.624958 master-0 kubenswrapper[4752]: I1205 10:35:18.624932 4752 container_manager_linux.go:303] "Creating device plugin manager" Dec 05 10:35:18.625068 master-0 kubenswrapper[4752]: I1205 10:35:18.625043 4752 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 10:35:18.625097 master-0 kubenswrapper[4752]: I1205 10:35:18.625079 4752 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 10:35:18.625231 master-0 kubenswrapper[4752]: I1205 10:35:18.625207 4752 state_mem.go:36] "Initialized new in-memory state store" Dec 05 10:35:18.625655 master-0 kubenswrapper[4752]: I1205 10:35:18.625627 4752 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 05 10:35:18.626414 master-0 kubenswrapper[4752]: I1205 10:35:18.626389 4752 kubelet.go:418] "Attempting to sync node with API server" Dec 05 10:35:18.626414 master-0 kubenswrapper[4752]: I1205 10:35:18.626411 4752 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 05 10:35:18.626532 master-0 kubenswrapper[4752]: I1205 10:35:18.626505 4752 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 05 10:35:18.626532 master-0 kubenswrapper[4752]: I1205 10:35:18.626528 4752 kubelet.go:324] "Adding apiserver pod source" Dec 05 10:35:18.626602 master-0 kubenswrapper[4752]: I1205 10:35:18.626545 4752 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 05 10:35:18.628482 master-0 kubenswrapper[4752]: I1205 10:35:18.628455 4752 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.13-2.rhaos4.18.git15789b8.el9" apiVersion="v1" Dec 05 10:35:18.629346 master-0 kubenswrapper[4752]: W1205 10:35:18.629218 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.sno.openstack.lab:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:18.629401 master-0 kubenswrapper[4752]: W1205 10:35:18.629222 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.sno.openstack.lab:6443/api/v1/nodes?fieldSelector=metadata.name%3Dmaster-0&limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:18.629565 master-0 kubenswrapper[4752]: E1205 10:35:18.629415 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes?fieldSelector=metadata.name%3Dmaster-0&limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:18.629624 master-0 kubenswrapper[4752]: E1205 10:35:18.629397 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.sno.openstack.lab:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:18.629624 master-0 kubenswrapper[4752]: I1205 10:35:18.629605 4752 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 05 10:35:18.629882 master-0 kubenswrapper[4752]: I1205 10:35:18.629850 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 05 10:35:18.629924 master-0 kubenswrapper[4752]: I1205 10:35:18.629884 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 05 10:35:18.629924 master-0 kubenswrapper[4752]: I1205 10:35:18.629894 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 05 10:35:18.629924 master-0 kubenswrapper[4752]: I1205 10:35:18.629900 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 05 10:35:18.629924 master-0 kubenswrapper[4752]: I1205 10:35:18.629907 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 05 10:35:18.629924 master-0 kubenswrapper[4752]: I1205 10:35:18.629914 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 05 10:35:18.629924 master-0 kubenswrapper[4752]: I1205 10:35:18.629921 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 05 10:35:18.629924 master-0 kubenswrapper[4752]: I1205 10:35:18.629928 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 05 10:35:18.630123 master-0 kubenswrapper[4752]: I1205 10:35:18.629939 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 05 10:35:18.630123 master-0 kubenswrapper[4752]: I1205 10:35:18.629957 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 05 10:35:18.630123 master-0 kubenswrapper[4752]: I1205 10:35:18.629977 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 05 10:35:18.630123 master-0 kubenswrapper[4752]: I1205 10:35:18.629992 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 05 10:35:18.630456 master-0 kubenswrapper[4752]: I1205 10:35:18.630394 4752 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 05 10:35:18.630818 master-0 kubenswrapper[4752]: I1205 10:35:18.630787 4752 server.go:1280] "Started kubelet" Dec 05 10:35:18.631122 master-0 kubenswrapper[4752]: I1205 10:35:18.631044 4752 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 05 10:35:18.631912 master-0 systemd[1]: Started Kubernetes Kubelet. Dec 05 10:35:18.632215 master-0 kubenswrapper[4752]: I1205 10:35:18.631508 4752 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 05 10:35:18.632870 master-0 kubenswrapper[4752]: I1205 10:35:18.632630 4752 server_v1.go:47] "podresources" method="list" useActivePods=true Dec 05 10:35:18.633565 master-0 kubenswrapper[4752]: I1205 10:35:18.633439 4752 server.go:449] "Adding debug handlers to kubelet server" Dec 05 10:35:18.635293 master-0 kubenswrapper[4752]: I1205 10:35:18.635011 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:18.636260 master-0 kubenswrapper[4752]: I1205 10:35:18.635486 4752 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 05 10:35:18.636260 master-0 kubenswrapper[4752]: I1205 10:35:18.635840 4752 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 05 10:35:18.636700 master-0 kubenswrapper[4752]: I1205 10:35:18.636624 4752 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 05 10:35:18.636838 master-0 kubenswrapper[4752]: E1205 10:35:18.636798 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:18.636900 master-0 kubenswrapper[4752]: I1205 10:35:18.636830 4752 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Dec 05 10:35:18.636900 master-0 kubenswrapper[4752]: I1205 10:35:18.636813 4752 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 05 10:35:18.636900 master-0 kubenswrapper[4752]: I1205 10:35:18.636892 4752 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 05 10:35:18.639290 master-0 kubenswrapper[4752]: I1205 10:35:18.639084 4752 reconstruct.go:97] "Volume reconstruction finished" Dec 05 10:35:18.640054 master-0 kubenswrapper[4752]: I1205 10:35:18.639252 4752 reconciler.go:26] "Reconciler: start to sync state" Dec 05 10:35:18.640356 master-0 kubenswrapper[4752]: E1205 10:35:18.640145 4752 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="200ms" Dec 05 10:35:18.640877 master-0 kubenswrapper[4752]: W1205 10:35:18.640789 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:18.640877 master-0 kubenswrapper[4752]: E1205 10:35:18.640864 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:18.644817 master-0 kubenswrapper[4752]: I1205 10:35:18.644776 4752 factory.go:55] Registering systemd factory Dec 05 10:35:18.644817 master-0 kubenswrapper[4752]: I1205 10:35:18.644804 4752 factory.go:221] Registration of the systemd container factory successfully Dec 05 10:35:18.644924 master-0 kubenswrapper[4752]: E1205 10:35:18.644885 4752 kubelet.go:1495] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Dec 05 10:35:18.645364 master-0 kubenswrapper[4752]: I1205 10:35:18.645334 4752 factory.go:153] Registering CRI-O factory Dec 05 10:35:18.645364 master-0 kubenswrapper[4752]: I1205 10:35:18.645362 4752 factory.go:221] Registration of the crio container factory successfully Dec 05 10:35:18.645490 master-0 kubenswrapper[4752]: I1205 10:35:18.645468 4752 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 05 10:35:18.645545 master-0 kubenswrapper[4752]: I1205 10:35:18.645512 4752 factory.go:103] Registering Raw factory Dec 05 10:35:18.645545 master-0 kubenswrapper[4752]: I1205 10:35:18.645534 4752 manager.go:1196] Started watching for new ooms in manager Dec 05 10:35:18.646107 master-0 kubenswrapper[4752]: E1205 10:35:18.645626 4752 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/default/events\": dial tcp 192.168.32.10:6443: connect: connection refused" event="&Event{ObjectMeta:{master-0.187e4b51ca172981 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.630762881 +0000 UTC m=+0.170886210,LastTimestamp:2025-12-05 10:35:18.630762881 +0000 UTC m=+0.170886210,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:18.646281 master-0 kubenswrapper[4752]: I1205 10:35:18.646259 4752 manager.go:319] Starting recovery of all containers Dec 05 10:35:18.668573 master-0 kubenswrapper[4752]: I1205 10:35:18.668320 4752 manager.go:324] Recovery completed Dec 05 10:35:18.682970 master-0 kubenswrapper[4752]: I1205 10:35:18.682911 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.684630 master-0 kubenswrapper[4752]: I1205 10:35:18.684591 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.684730 master-0 kubenswrapper[4752]: I1205 10:35:18.684639 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.684730 master-0 kubenswrapper[4752]: I1205 10:35:18.684648 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.685349 master-0 kubenswrapper[4752]: I1205 10:35:18.685298 4752 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 05 10:35:18.685349 master-0 kubenswrapper[4752]: I1205 10:35:18.685344 4752 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 05 10:35:18.685488 master-0 kubenswrapper[4752]: I1205 10:35:18.685377 4752 state_mem.go:36] "Initialized new in-memory state store" Dec 05 10:35:18.693870 master-0 kubenswrapper[4752]: I1205 10:35:18.693821 4752 policy_none.go:49] "None policy: Start" Dec 05 10:35:18.694734 master-0 kubenswrapper[4752]: I1205 10:35:18.694700 4752 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 05 10:35:18.694734 master-0 kubenswrapper[4752]: I1205 10:35:18.694725 4752 state_mem.go:35] "Initializing new in-memory state store" Dec 05 10:35:18.737683 master-0 kubenswrapper[4752]: E1205 10:35:18.737639 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:18.787070 master-0 kubenswrapper[4752]: I1205 10:35:18.752747 4752 manager.go:334] "Starting Device Plugin manager" Dec 05 10:35:18.787070 master-0 kubenswrapper[4752]: I1205 10:35:18.752802 4752 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 05 10:35:18.787070 master-0 kubenswrapper[4752]: I1205 10:35:18.752814 4752 server.go:79] "Starting device plugin registration server" Dec 05 10:35:18.787070 master-0 kubenswrapper[4752]: I1205 10:35:18.753409 4752 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 05 10:35:18.787070 master-0 kubenswrapper[4752]: I1205 10:35:18.753437 4752 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 05 10:35:18.787070 master-0 kubenswrapper[4752]: I1205 10:35:18.753558 4752 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 05 10:35:18.787070 master-0 kubenswrapper[4752]: I1205 10:35:18.753673 4752 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 05 10:35:18.787070 master-0 kubenswrapper[4752]: I1205 10:35:18.753686 4752 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 05 10:35:18.787070 master-0 kubenswrapper[4752]: E1205 10:35:18.755182 4752 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"master-0\" not found" Dec 05 10:35:18.793162 master-0 kubenswrapper[4752]: I1205 10:35:18.793103 4752 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 05 10:35:18.794828 master-0 kubenswrapper[4752]: I1205 10:35:18.794622 4752 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 05 10:35:18.794828 master-0 kubenswrapper[4752]: I1205 10:35:18.794688 4752 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 05 10:35:18.794828 master-0 kubenswrapper[4752]: I1205 10:35:18.794734 4752 kubelet.go:2335] "Starting kubelet main sync loop" Dec 05 10:35:18.794828 master-0 kubenswrapper[4752]: E1205 10:35:18.794793 4752 kubelet.go:2359] "Skipping pod synchronization" err="PLEG is not healthy: pleg has yet to be successful" Dec 05 10:35:18.796286 master-0 kubenswrapper[4752]: W1205 10:35:18.796181 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.sno.openstack.lab:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:18.796353 master-0 kubenswrapper[4752]: E1205 10:35:18.796309 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.sno.openstack.lab:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:18.841536 master-0 kubenswrapper[4752]: E1205 10:35:18.841450 4752 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="400ms" Dec 05 10:35:18.854593 master-0 kubenswrapper[4752]: I1205 10:35:18.854526 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.855448 master-0 kubenswrapper[4752]: I1205 10:35:18.855395 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.855448 master-0 kubenswrapper[4752]: I1205 10:35:18.855445 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.855448 master-0 kubenswrapper[4752]: I1205 10:35:18.855454 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.855566 master-0 kubenswrapper[4752]: I1205 10:35:18.855479 4752 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:35:18.856361 master-0 kubenswrapper[4752]: E1205 10:35:18.856293 4752 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/nodes\": dial tcp 192.168.32.10:6443: connect: connection refused" node="master-0" Dec 05 10:35:18.895630 master-0 kubenswrapper[4752]: I1205 10:35:18.895450 4752 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-master-0","openshift-etcd/etcd-master-0-master-0","openshift-kube-apiserver/bootstrap-kube-apiserver-master-0","kube-system/bootstrap-kube-controller-manager-master-0","kube-system/bootstrap-kube-scheduler-master-0"] Dec 05 10:35:18.895630 master-0 kubenswrapper[4752]: I1205 10:35:18.895545 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.896772 master-0 kubenswrapper[4752]: I1205 10:35:18.896706 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.896772 master-0 kubenswrapper[4752]: I1205 10:35:18.896759 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.896772 master-0 kubenswrapper[4752]: I1205 10:35:18.896768 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.896905 master-0 kubenswrapper[4752]: I1205 10:35:18.896883 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.897167 master-0 kubenswrapper[4752]: I1205 10:35:18.897130 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:35:18.897203 master-0 kubenswrapper[4752]: I1205 10:35:18.897178 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.897839 master-0 kubenswrapper[4752]: I1205 10:35:18.897777 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.897839 master-0 kubenswrapper[4752]: I1205 10:35:18.897816 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.897839 master-0 kubenswrapper[4752]: I1205 10:35:18.897830 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.898267 master-0 kubenswrapper[4752]: I1205 10:35:18.897898 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.898267 master-0 kubenswrapper[4752]: I1205 10:35:18.897917 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.898267 master-0 kubenswrapper[4752]: I1205 10:35:18.897932 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.898267 master-0 kubenswrapper[4752]: I1205 10:35:18.897947 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.898267 master-0 kubenswrapper[4752]: I1205 10:35:18.898037 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:35:18.898267 master-0 kubenswrapper[4752]: I1205 10:35:18.898059 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.898957 master-0 kubenswrapper[4752]: I1205 10:35:18.898741 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.898957 master-0 kubenswrapper[4752]: I1205 10:35:18.898763 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.898957 master-0 kubenswrapper[4752]: I1205 10:35:18.898771 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.898957 master-0 kubenswrapper[4752]: I1205 10:35:18.898832 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.898957 master-0 kubenswrapper[4752]: I1205 10:35:18.898849 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.898957 master-0 kubenswrapper[4752]: I1205 10:35:18.898860 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.899353 master-0 kubenswrapper[4752]: I1205 10:35:18.899068 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.899353 master-0 kubenswrapper[4752]: I1205 10:35:18.899181 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:18.899353 master-0 kubenswrapper[4752]: I1205 10:35:18.899205 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.899818 master-0 kubenswrapper[4752]: I1205 10:35:18.899770 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.899818 master-0 kubenswrapper[4752]: I1205 10:35:18.899797 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.899818 master-0 kubenswrapper[4752]: I1205 10:35:18.899810 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.900101 master-0 kubenswrapper[4752]: I1205 10:35:18.899888 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.900101 master-0 kubenswrapper[4752]: I1205 10:35:18.899812 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.900101 master-0 kubenswrapper[4752]: I1205 10:35:18.899956 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.900101 master-0 kubenswrapper[4752]: I1205 10:35:18.899968 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.900603 master-0 kubenswrapper[4752]: I1205 10:35:18.900556 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.900603 master-0 kubenswrapper[4752]: I1205 10:35:18.900584 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.900603 master-0 kubenswrapper[4752]: I1205 10:35:18.900595 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.900890 master-0 kubenswrapper[4752]: I1205 10:35:18.900561 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:18.900890 master-0 kubenswrapper[4752]: I1205 10:35:18.900665 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.900890 master-0 kubenswrapper[4752]: I1205 10:35:18.900716 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:35:18.900890 master-0 kubenswrapper[4752]: I1205 10:35:18.900738 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:18.901469 master-0 kubenswrapper[4752]: I1205 10:35:18.901381 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.901469 master-0 kubenswrapper[4752]: I1205 10:35:18.901402 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.901469 master-0 kubenswrapper[4752]: I1205 10:35:18.901410 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.901752 master-0 kubenswrapper[4752]: I1205 10:35:18.901519 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:18.901752 master-0 kubenswrapper[4752]: I1205 10:35:18.901529 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:18.901752 master-0 kubenswrapper[4752]: I1205 10:35:18.901536 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:18.941376 master-0 kubenswrapper[4752]: I1205 10:35:18.941311 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-secrets\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:35:18.941376 master-0 kubenswrapper[4752]: I1205 10:35:18.941357 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-etc-kube\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:35:18.941376 master-0 kubenswrapper[4752]: I1205 10:35:18.941385 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-ssl-certs-host\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:18.941654 master-0 kubenswrapper[4752]: I1205 10:35:18.941460 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-config\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:18.941654 master-0 kubenswrapper[4752]: I1205 10:35:18.941504 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-logs\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:18.941654 master-0 kubenswrapper[4752]: I1205 10:35:18.941529 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-logs\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:35:18.941654 master-0 kubenswrapper[4752]: I1205 10:35:18.941546 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-certs\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:35:18.941654 master-0 kubenswrapper[4752]: I1205 10:35:18.941586 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-secrets\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:18.941654 master-0 kubenswrapper[4752]: I1205 10:35:18.941626 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-logs\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:18.941853 master-0 kubenswrapper[4752]: I1205 10:35:18.941662 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-secrets\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:18.941853 master-0 kubenswrapper[4752]: I1205 10:35:18.941683 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-etc-kubernetes-cloud\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:18.941853 master-0 kubenswrapper[4752]: I1205 10:35:18.941740 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-ssl-certs-host\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:18.941853 master-0 kubenswrapper[4752]: I1205 10:35:18.941759 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:35:18.941853 master-0 kubenswrapper[4752]: I1205 10:35:18.941775 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-data-dir\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:35:18.941853 master-0 kubenswrapper[4752]: I1205 10:35:18.941791 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-etc-kubernetes-cloud\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:18.942010 master-0 kubenswrapper[4752]: I1205 10:35:18.941878 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-audit-dir\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:18.942010 master-0 kubenswrapper[4752]: I1205 10:35:18.941952 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-config\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.042857 master-0 kubenswrapper[4752]: I1205 10:35:19.042761 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-config\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.043204 master-0 kubenswrapper[4752]: I1205 10:35:19.043138 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-secrets\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:35:19.043275 master-0 kubenswrapper[4752]: I1205 10:35:19.043230 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-config\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.043275 master-0 kubenswrapper[4752]: I1205 10:35:19.043245 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-secrets\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:35:19.043396 master-0 kubenswrapper[4752]: I1205 10:35:19.043308 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-etc-kube\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:35:19.043492 master-0 kubenswrapper[4752]: I1205 10:35:19.043245 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-etc-kube\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:35:19.043674 master-0 kubenswrapper[4752]: I1205 10:35:19.043597 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-ssl-certs-host\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.043727 master-0 kubenswrapper[4752]: I1205 10:35:19.043676 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-ssl-certs-host\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.043782 master-0 kubenswrapper[4752]: I1205 10:35:19.043744 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-config\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:19.043782 master-0 kubenswrapper[4752]: I1205 10:35:19.043774 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-config\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:19.043894 master-0 kubenswrapper[4752]: I1205 10:35:19.043842 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-logs\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:19.043894 master-0 kubenswrapper[4752]: I1205 10:35:19.043863 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-logs\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:19.043976 master-0 kubenswrapper[4752]: I1205 10:35:19.043955 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-logs\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:35:19.044095 master-0 kubenswrapper[4752]: I1205 10:35:19.044051 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-certs\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:35:19.044146 master-0 kubenswrapper[4752]: I1205 10:35:19.044092 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-logs\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:35:19.044146 master-0 kubenswrapper[4752]: I1205 10:35:19.044131 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-certs\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:35:19.044304 master-0 kubenswrapper[4752]: I1205 10:35:19.044263 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-secrets\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.044369 master-0 kubenswrapper[4752]: I1205 10:35:19.044347 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-logs\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.044484 master-0 kubenswrapper[4752]: I1205 10:35:19.044386 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-secrets\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:19.044529 master-0 kubenswrapper[4752]: I1205 10:35:19.044494 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-logs\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.044568 master-0 kubenswrapper[4752]: I1205 10:35:19.044456 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-secrets\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.044568 master-0 kubenswrapper[4752]: I1205 10:35:19.044434 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-secrets\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:19.044656 master-0 kubenswrapper[4752]: I1205 10:35:19.044520 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-etc-kubernetes-cloud\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:19.044656 master-0 kubenswrapper[4752]: I1205 10:35:19.044592 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-etc-kubernetes-cloud\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:19.044733 master-0 kubenswrapper[4752]: I1205 10:35:19.044626 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-ssl-certs-host\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:19.044800 master-0 kubenswrapper[4752]: I1205 10:35:19.044724 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:35:19.044900 master-0 kubenswrapper[4752]: I1205 10:35:19.044859 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-ssl-certs-host\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:19.044986 master-0 kubenswrapper[4752]: I1205 10:35:19.044957 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-data-dir\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:35:19.045072 master-0 kubenswrapper[4752]: I1205 10:35:19.045024 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-etc-kubernetes-cloud\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.045118 master-0 kubenswrapper[4752]: I1205 10:35:19.045085 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-etc-kubernetes-cloud\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.045118 master-0 kubenswrapper[4752]: I1205 10:35:19.044924 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:35:19.045202 master-0 kubenswrapper[4752]: I1205 10:35:19.045106 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-audit-dir\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.045202 master-0 kubenswrapper[4752]: I1205 10:35:19.045082 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-data-dir\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:35:19.045283 master-0 kubenswrapper[4752]: I1205 10:35:19.045175 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-audit-dir\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.056976 master-0 kubenswrapper[4752]: I1205 10:35:19.056901 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:19.058356 master-0 kubenswrapper[4752]: I1205 10:35:19.058305 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:19.058454 master-0 kubenswrapper[4752]: I1205 10:35:19.058359 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:19.058454 master-0 kubenswrapper[4752]: I1205 10:35:19.058377 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:19.058565 master-0 kubenswrapper[4752]: I1205 10:35:19.058472 4752 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:35:19.059701 master-0 kubenswrapper[4752]: E1205 10:35:19.059644 4752 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/nodes\": dial tcp 192.168.32.10:6443: connect: connection refused" node="master-0" Dec 05 10:35:19.243777 master-0 kubenswrapper[4752]: E1205 10:35:19.243501 4752 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="800ms" Dec 05 10:35:19.248277 master-0 kubenswrapper[4752]: I1205 10:35:19.248180 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:35:19.269625 master-0 kubenswrapper[4752]: I1205 10:35:19.269578 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:35:19.295797 master-0 kubenswrapper[4752]: I1205 10:35:19.295688 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:19.313180 master-0 kubenswrapper[4752]: I1205 10:35:19.313131 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:35:19.322195 master-0 kubenswrapper[4752]: I1205 10:35:19.322103 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:19.460401 master-0 kubenswrapper[4752]: I1205 10:35:19.460323 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:19.461764 master-0 kubenswrapper[4752]: I1205 10:35:19.461643 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:19.461764 master-0 kubenswrapper[4752]: I1205 10:35:19.461746 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:19.461764 master-0 kubenswrapper[4752]: I1205 10:35:19.461766 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:19.462317 master-0 kubenswrapper[4752]: I1205 10:35:19.461837 4752 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:35:19.463125 master-0 kubenswrapper[4752]: E1205 10:35:19.463009 4752 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/nodes\": dial tcp 192.168.32.10:6443: connect: connection refused" node="master-0" Dec 05 10:35:19.465912 master-0 kubenswrapper[4752]: W1205 10:35:19.465789 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.sno.openstack.lab:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:19.466053 master-0 kubenswrapper[4752]: E1205 10:35:19.465932 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.sno.openstack.lab:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:19.602026 master-0 kubenswrapper[4752]: W1205 10:35:19.601937 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:19.602026 master-0 kubenswrapper[4752]: E1205 10:35:19.602017 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:19.638083 master-0 kubenswrapper[4752]: I1205 10:35:19.637896 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:19.845135 master-0 kubenswrapper[4752]: W1205 10:35:19.845060 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3169f44496ed8a28c6d6a15511ab0eec.slice/crio-d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7 WatchSource:0}: Error finding container d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7: Status 404 returned error can't find the container with id d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7 Dec 05 10:35:19.859761 master-0 kubenswrapper[4752]: I1205 10:35:19.859709 4752 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:35:19.862615 master-0 kubenswrapper[4752]: W1205 10:35:19.862563 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.sno.openstack.lab:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:19.862732 master-0 kubenswrapper[4752]: E1205 10:35:19.862616 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.sno.openstack.lab:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:19.874369 master-0 kubenswrapper[4752]: W1205 10:35:19.874315 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75143d9bc4a2dc15781dc51ccff632a.slice/crio-09d6c752fa5b21d6de28695ca1e30252206036602e7b7cad7c82ecf0c820105d WatchSource:0}: Error finding container 09d6c752fa5b21d6de28695ca1e30252206036602e7b7cad7c82ecf0c820105d: Status 404 returned error can't find the container with id 09d6c752fa5b21d6de28695ca1e30252206036602e7b7cad7c82ecf0c820105d Dec 05 10:35:19.907738 master-0 kubenswrapper[4752]: W1205 10:35:19.907666 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e09e2af7200e6f9be469dbfd9bb1127.slice/crio-611473a6bb698264f5594083fb6219f23f5848bd85b0fdd8dd5932b32fae3c36 WatchSource:0}: Error finding container 611473a6bb698264f5594083fb6219f23f5848bd85b0fdd8dd5932b32fae3c36: Status 404 returned error can't find the container with id 611473a6bb698264f5594083fb6219f23f5848bd85b0fdd8dd5932b32fae3c36 Dec 05 10:35:19.917843 master-0 kubenswrapper[4752]: W1205 10:35:19.917770 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b47694fcc32464ab24d09c23d6efb57.slice/crio-7672f24a6f45e585db5dc690d30bd20c69d755aa3cc9d4346ea6ca7923305116 WatchSource:0}: Error finding container 7672f24a6f45e585db5dc690d30bd20c69d755aa3cc9d4346ea6ca7923305116: Status 404 returned error can't find the container with id 7672f24a6f45e585db5dc690d30bd20c69d755aa3cc9d4346ea6ca7923305116 Dec 05 10:35:19.937309 master-0 kubenswrapper[4752]: W1205 10:35:19.937251 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc0396a9a2689b3e8c132c12640cbe83.slice/crio-87b1749af9a101502a0aa9cd2d75f0e1562789f11a001117a5d269d9ed620723 WatchSource:0}: Error finding container 87b1749af9a101502a0aa9cd2d75f0e1562789f11a001117a5d269d9ed620723: Status 404 returned error can't find the container with id 87b1749af9a101502a0aa9cd2d75f0e1562789f11a001117a5d269d9ed620723 Dec 05 10:35:20.045857 master-0 kubenswrapper[4752]: E1205 10:35:20.045715 4752 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="1.6s" Dec 05 10:35:20.164202 master-0 kubenswrapper[4752]: W1205 10:35:20.163963 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.sno.openstack.lab:6443/api/v1/nodes?fieldSelector=metadata.name%3Dmaster-0&limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:20.164202 master-0 kubenswrapper[4752]: E1205 10:35:20.164083 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes?fieldSelector=metadata.name%3Dmaster-0&limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:20.263359 master-0 kubenswrapper[4752]: I1205 10:35:20.263253 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:20.267159 master-0 kubenswrapper[4752]: I1205 10:35:20.267083 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:20.267159 master-0 kubenswrapper[4752]: I1205 10:35:20.267160 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:20.267338 master-0 kubenswrapper[4752]: I1205 10:35:20.267188 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:20.267338 master-0 kubenswrapper[4752]: I1205 10:35:20.267292 4752 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:35:20.269288 master-0 kubenswrapper[4752]: E1205 10:35:20.269031 4752 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/nodes\": dial tcp 192.168.32.10:6443: connect: connection refused" node="master-0" Dec 05 10:35:20.626988 master-0 kubenswrapper[4752]: I1205 10:35:20.626915 4752 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 05 10:35:20.628156 master-0 kubenswrapper[4752]: E1205 10:35:20.628109 4752 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.sno.openstack.lab:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:20.637241 master-0 kubenswrapper[4752]: I1205 10:35:20.637202 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:20.801745 master-0 kubenswrapper[4752]: I1205 10:35:20.801607 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" event={"ID":"d75143d9bc4a2dc15781dc51ccff632a","Type":"ContainerStarted","Data":"09d6c752fa5b21d6de28695ca1e30252206036602e7b7cad7c82ecf0c820105d"} Dec 05 10:35:20.802480 master-0 kubenswrapper[4752]: I1205 10:35:20.802451 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerStarted","Data":"d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7"} Dec 05 10:35:20.803386 master-0 kubenswrapper[4752]: I1205 10:35:20.803321 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0-master-0" event={"ID":"cc0396a9a2689b3e8c132c12640cbe83","Type":"ContainerStarted","Data":"87b1749af9a101502a0aa9cd2d75f0e1562789f11a001117a5d269d9ed620723"} Dec 05 10:35:20.804536 master-0 kubenswrapper[4752]: I1205 10:35:20.804469 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerStarted","Data":"7672f24a6f45e585db5dc690d30bd20c69d755aa3cc9d4346ea6ca7923305116"} Dec 05 10:35:20.805919 master-0 kubenswrapper[4752]: I1205 10:35:20.805877 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-scheduler-master-0" event={"ID":"5e09e2af7200e6f9be469dbfd9bb1127","Type":"ContainerStarted","Data":"611473a6bb698264f5594083fb6219f23f5848bd85b0fdd8dd5932b32fae3c36"} Dec 05 10:35:21.545927 master-0 kubenswrapper[4752]: W1205 10:35:21.545862 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.sno.openstack.lab:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:21.546117 master-0 kubenswrapper[4752]: E1205 10:35:21.545952 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.sno.openstack.lab:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:21.637004 master-0 kubenswrapper[4752]: I1205 10:35:21.636932 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:21.646804 master-0 kubenswrapper[4752]: E1205 10:35:21.646755 4752 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="3.2s" Dec 05 10:35:21.870061 master-0 kubenswrapper[4752]: I1205 10:35:21.869921 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:21.871387 master-0 kubenswrapper[4752]: I1205 10:35:21.871363 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:21.871467 master-0 kubenswrapper[4752]: I1205 10:35:21.871403 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:21.871467 master-0 kubenswrapper[4752]: I1205 10:35:21.871443 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:21.871559 master-0 kubenswrapper[4752]: I1205 10:35:21.871513 4752 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:35:21.872320 master-0 kubenswrapper[4752]: E1205 10:35:21.872276 4752 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/nodes\": dial tcp 192.168.32.10:6443: connect: connection refused" node="master-0" Dec 05 10:35:21.952710 master-0 kubenswrapper[4752]: W1205 10:35:21.952625 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:21.952896 master-0 kubenswrapper[4752]: E1205 10:35:21.952730 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:22.321736 master-0 kubenswrapper[4752]: W1205 10:35:22.321667 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.sno.openstack.lab:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:22.321955 master-0 kubenswrapper[4752]: E1205 10:35:22.321758 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.sno.openstack.lab:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:22.638114 master-0 kubenswrapper[4752]: I1205 10:35:22.637957 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:23.359994 master-0 kubenswrapper[4752]: W1205 10:35:23.359713 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.sno.openstack.lab:6443/api/v1/nodes?fieldSelector=metadata.name%3Dmaster-0&limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:23.359994 master-0 kubenswrapper[4752]: E1205 10:35:23.359991 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes?fieldSelector=metadata.name%3Dmaster-0&limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:23.637609 master-0 kubenswrapper[4752]: I1205 10:35:23.637556 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:23.812862 master-0 kubenswrapper[4752]: I1205 10:35:23.812811 4752 generic.go:334] "Generic (PLEG): container finished" podID="3169f44496ed8a28c6d6a15511ab0eec" containerID="2a995b4c36846fd814a8f2ca3bd01eb60eaabd9a004a171a036b25b030f63eff" exitCode=0 Dec 05 10:35:23.813685 master-0 kubenswrapper[4752]: I1205 10:35:23.812881 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerDied","Data":"2a995b4c36846fd814a8f2ca3bd01eb60eaabd9a004a171a036b25b030f63eff"} Dec 05 10:35:23.813685 master-0 kubenswrapper[4752]: I1205 10:35:23.812973 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:23.813806 master-0 kubenswrapper[4752]: I1205 10:35:23.813760 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:23.813806 master-0 kubenswrapper[4752]: I1205 10:35:23.813779 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:23.813806 master-0 kubenswrapper[4752]: I1205 10:35:23.813788 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:23.815844 master-0 kubenswrapper[4752]: I1205 10:35:23.815791 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0-master-0" event={"ID":"cc0396a9a2689b3e8c132c12640cbe83","Type":"ContainerStarted","Data":"99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c"} Dec 05 10:35:23.815844 master-0 kubenswrapper[4752]: I1205 10:35:23.815821 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:23.815844 master-0 kubenswrapper[4752]: I1205 10:35:23.815843 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0-master-0" event={"ID":"cc0396a9a2689b3e8c132c12640cbe83","Type":"ContainerStarted","Data":"619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b"} Dec 05 10:35:23.816491 master-0 kubenswrapper[4752]: I1205 10:35:23.816459 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:23.816557 master-0 kubenswrapper[4752]: I1205 10:35:23.816494 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:23.816557 master-0 kubenswrapper[4752]: I1205 10:35:23.816506 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:24.639790 master-0 kubenswrapper[4752]: I1205 10:35:24.639740 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:24.774086 master-0 kubenswrapper[4752]: I1205 10:35:24.773932 4752 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 05 10:35:24.775612 master-0 kubenswrapper[4752]: E1205 10:35:24.775565 4752 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.sno.openstack.lab:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:24.819958 master-0 kubenswrapper[4752]: I1205 10:35:24.819901 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_kube-rbac-proxy-crio-master-0_3169f44496ed8a28c6d6a15511ab0eec/kube-rbac-proxy-crio/0.log" Dec 05 10:35:24.820726 master-0 kubenswrapper[4752]: I1205 10:35:24.820358 4752 generic.go:334] "Generic (PLEG): container finished" podID="3169f44496ed8a28c6d6a15511ab0eec" containerID="8a19d2608ebc4e5c05a6953726ca619064758c7c9878bedda59471107f22e937" exitCode=1 Dec 05 10:35:24.820726 master-0 kubenswrapper[4752]: I1205 10:35:24.820463 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:24.820866 master-0 kubenswrapper[4752]: I1205 10:35:24.820842 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:24.821085 master-0 kubenswrapper[4752]: I1205 10:35:24.821006 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerDied","Data":"8a19d2608ebc4e5c05a6953726ca619064758c7c9878bedda59471107f22e937"} Dec 05 10:35:24.822385 master-0 kubenswrapper[4752]: I1205 10:35:24.821453 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:24.822385 master-0 kubenswrapper[4752]: I1205 10:35:24.821511 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:24.822385 master-0 kubenswrapper[4752]: I1205 10:35:24.821530 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:24.822385 master-0 kubenswrapper[4752]: I1205 10:35:24.822063 4752 scope.go:117] "RemoveContainer" containerID="8a19d2608ebc4e5c05a6953726ca619064758c7c9878bedda59471107f22e937" Dec 05 10:35:24.823073 master-0 kubenswrapper[4752]: I1205 10:35:24.822947 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:24.823073 master-0 kubenswrapper[4752]: I1205 10:35:24.822980 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:24.823073 master-0 kubenswrapper[4752]: I1205 10:35:24.822992 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:24.848219 master-0 kubenswrapper[4752]: E1205 10:35:24.848175 4752 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="6.4s" Dec 05 10:35:25.073077 master-0 kubenswrapper[4752]: I1205 10:35:25.073042 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:25.074082 master-0 kubenswrapper[4752]: I1205 10:35:25.074053 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:25.074153 master-0 kubenswrapper[4752]: I1205 10:35:25.074099 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:25.074153 master-0 kubenswrapper[4752]: I1205 10:35:25.074110 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:25.074209 master-0 kubenswrapper[4752]: I1205 10:35:25.074167 4752 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:35:25.074924 master-0 kubenswrapper[4752]: E1205 10:35:25.074882 4752 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/nodes\": dial tcp 192.168.32.10:6443: connect: connection refused" node="master-0" Dec 05 10:35:25.336273 master-0 kubenswrapper[4752]: W1205 10:35:25.336116 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:25.336273 master-0 kubenswrapper[4752]: E1205 10:35:25.336209 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:25.363469 master-0 kubenswrapper[4752]: E1205 10:35:25.363141 4752 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/default/events\": dial tcp 192.168.32.10:6443: connect: connection refused" event="&Event{ObjectMeta:{master-0.187e4b51ca172981 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.630762881 +0000 UTC m=+0.170886210,LastTimestamp:2025-12-05 10:35:18.630762881 +0000 UTC m=+0.170886210,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:25.638134 master-0 kubenswrapper[4752]: I1205 10:35:25.638024 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:26.638395 master-0 kubenswrapper[4752]: I1205 10:35:26.638337 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:27.341983 master-0 kubenswrapper[4752]: W1205 10:35:27.341703 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.sno.openstack.lab:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:27.341983 master-0 kubenswrapper[4752]: E1205 10:35:27.341790 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.sno.openstack.lab:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:27.637544 master-0 kubenswrapper[4752]: I1205 10:35:27.637313 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:27.784547 master-0 kubenswrapper[4752]: W1205 10:35:27.784464 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.sno.openstack.lab:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:27.784547 master-0 kubenswrapper[4752]: E1205 10:35:27.784530 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.sno.openstack.lab:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:28.639045 master-0 kubenswrapper[4752]: I1205 10:35:28.638914 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:28.755375 master-0 kubenswrapper[4752]: E1205 10:35:28.755279 4752 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"master-0\" not found" Dec 05 10:35:29.214130 master-0 kubenswrapper[4752]: W1205 10:35:29.214081 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.sno.openstack.lab:6443/api/v1/nodes?fieldSelector=metadata.name%3Dmaster-0&limit=500&resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:29.214497 master-0 kubenswrapper[4752]: E1205 10:35:29.214141 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes?fieldSelector=metadata.name%3Dmaster-0&limit=500&resourceVersion=0\": dial tcp 192.168.32.10:6443: connect: connection refused" logger="UnhandledError" Dec 05 10:35:29.637939 master-0 kubenswrapper[4752]: I1205 10:35:29.637866 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:29.833025 master-0 kubenswrapper[4752]: I1205 10:35:29.832972 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerStarted","Data":"6c524eb37d9eb519540115909748ea1e0272427fa38212f1b6de731f3f942edd"} Dec 05 10:35:29.835245 master-0 kubenswrapper[4752]: I1205 10:35:29.835214 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-scheduler-master-0" event={"ID":"5e09e2af7200e6f9be469dbfd9bb1127","Type":"ContainerStarted","Data":"f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782"} Dec 05 10:35:29.835388 master-0 kubenswrapper[4752]: I1205 10:35:29.835325 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:29.840004 master-0 kubenswrapper[4752]: I1205 10:35:29.839606 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:29.840004 master-0 kubenswrapper[4752]: I1205 10:35:29.839649 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:29.840004 master-0 kubenswrapper[4752]: I1205 10:35:29.839658 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:29.850440 master-0 kubenswrapper[4752]: I1205 10:35:29.850246 4752 generic.go:334] "Generic (PLEG): container finished" podID="d75143d9bc4a2dc15781dc51ccff632a" containerID="738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948" exitCode=0 Dec 05 10:35:29.850440 master-0 kubenswrapper[4752]: I1205 10:35:29.850318 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" event={"ID":"d75143d9bc4a2dc15781dc51ccff632a","Type":"ContainerDied","Data":"738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948"} Dec 05 10:35:29.850440 master-0 kubenswrapper[4752]: I1205 10:35:29.850349 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.850995 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.851031 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.851042 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.852266 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_kube-rbac-proxy-crio-master-0_3169f44496ed8a28c6d6a15511ab0eec/kube-rbac-proxy-crio/1.log" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.852609 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_kube-rbac-proxy-crio-master-0_3169f44496ed8a28c6d6a15511ab0eec/kube-rbac-proxy-crio/0.log" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.852923 4752 generic.go:334] "Generic (PLEG): container finished" podID="3169f44496ed8a28c6d6a15511ab0eec" containerID="aff810b496908f638cc4110ee6780958510ed5ee8e621db2ccc0eb661ea287e6" exitCode=1 Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.852957 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerDied","Data":"aff810b496908f638cc4110ee6780958510ed5ee8e621db2ccc0eb661ea287e6"} Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.853014 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.853021 4752 scope.go:117] "RemoveContainer" containerID="8a19d2608ebc4e5c05a6953726ca619064758c7c9878bedda59471107f22e937" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.853204 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.853620 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.853637 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.853646 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.854104 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.854145 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.854157 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: I1205 10:35:29.854606 4752 scope.go:117] "RemoveContainer" containerID="aff810b496908f638cc4110ee6780958510ed5ee8e621db2ccc0eb661ea287e6" Dec 05 10:35:29.855436 master-0 kubenswrapper[4752]: E1205 10:35:29.854789 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy-crio\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-rbac-proxy-crio pod=kube-rbac-proxy-crio-master-0_openshift-machine-config-operator(3169f44496ed8a28c6d6a15511ab0eec)\"" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" podUID="3169f44496ed8a28c6d6a15511ab0eec" Dec 05 10:35:30.637472 master-0 kubenswrapper[4752]: I1205 10:35:30.637391 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.sno.openstack.lab:6443/apis/storage.k8s.io/v1/csinodes/master-0?resourceVersion=0": dial tcp 192.168.32.10:6443: connect: connection refused Dec 05 10:35:30.857063 master-0 kubenswrapper[4752]: I1205 10:35:30.856991 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_kube-rbac-proxy-crio-master-0_3169f44496ed8a28c6d6a15511ab0eec/kube-rbac-proxy-crio/1.log" Dec 05 10:35:30.857492 master-0 kubenswrapper[4752]: I1205 10:35:30.857469 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:30.858271 master-0 kubenswrapper[4752]: I1205 10:35:30.858221 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:30.858341 master-0 kubenswrapper[4752]: I1205 10:35:30.858274 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:30.858341 master-0 kubenswrapper[4752]: I1205 10:35:30.858288 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:30.858712 master-0 kubenswrapper[4752]: I1205 10:35:30.858673 4752 generic.go:334] "Generic (PLEG): container finished" podID="8b47694fcc32464ab24d09c23d6efb57" containerID="6c524eb37d9eb519540115909748ea1e0272427fa38212f1b6de731f3f942edd" exitCode=1 Dec 05 10:35:30.858712 master-0 kubenswrapper[4752]: I1205 10:35:30.858702 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerDied","Data":"6c524eb37d9eb519540115909748ea1e0272427fa38212f1b6de731f3f942edd"} Dec 05 10:35:30.858712 master-0 kubenswrapper[4752]: I1205 10:35:30.858720 4752 scope.go:117] "RemoveContainer" containerID="aff810b496908f638cc4110ee6780958510ed5ee8e621db2ccc0eb661ea287e6" Dec 05 10:35:30.858975 master-0 kubenswrapper[4752]: E1205 10:35:30.858894 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy-crio\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-rbac-proxy-crio pod=kube-rbac-proxy-crio-master-0_openshift-machine-config-operator(3169f44496ed8a28c6d6a15511ab0eec)\"" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" podUID="3169f44496ed8a28c6d6a15511ab0eec" Dec 05 10:35:30.860275 master-0 kubenswrapper[4752]: I1205 10:35:30.860246 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" event={"ID":"d75143d9bc4a2dc15781dc51ccff632a","Type":"ContainerStarted","Data":"10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298"} Dec 05 10:35:30.860275 master-0 kubenswrapper[4752]: I1205 10:35:30.860271 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:30.860796 master-0 kubenswrapper[4752]: I1205 10:35:30.860773 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:30.860796 master-0 kubenswrapper[4752]: I1205 10:35:30.860798 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:30.860900 master-0 kubenswrapper[4752]: I1205 10:35:30.860807 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:31.475282 master-0 kubenswrapper[4752]: I1205 10:35:31.475198 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:31.476453 master-0 kubenswrapper[4752]: I1205 10:35:31.476358 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:31.476453 master-0 kubenswrapper[4752]: I1205 10:35:31.476443 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:31.476453 master-0 kubenswrapper[4752]: I1205 10:35:31.476453 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:31.476727 master-0 kubenswrapper[4752]: I1205 10:35:31.476503 4752 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:35:32.225773 master-0 kubenswrapper[4752]: E1205 10:35:32.225706 4752 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="master-0" Dec 05 10:35:32.227586 master-0 kubenswrapper[4752]: I1205 10:35:32.227547 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:32.228152 master-0 kubenswrapper[4752]: E1205 10:35:32.228119 4752 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"master-0\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Dec 05 10:35:32.638300 master-0 kubenswrapper[4752]: I1205 10:35:32.638271 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:32.816478 master-0 kubenswrapper[4752]: W1205 10:35:32.816436 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:32.816707 master-0 kubenswrapper[4752]: E1205 10:35:32.816484 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Dec 05 10:35:32.867088 master-0 kubenswrapper[4752]: I1205 10:35:32.866987 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerStarted","Data":"80aee350277cce308d84133aa5eefc6be3ed0712557f82d0d22b78dac448f0af"} Dec 05 10:35:32.867261 master-0 kubenswrapper[4752]: I1205 10:35:32.867118 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:32.868503 master-0 kubenswrapper[4752]: I1205 10:35:32.867884 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:32.868503 master-0 kubenswrapper[4752]: I1205 10:35:32.867919 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:32.868503 master-0 kubenswrapper[4752]: I1205 10:35:32.867933 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:32.868503 master-0 kubenswrapper[4752]: I1205 10:35:32.868218 4752 scope.go:117] "RemoveContainer" containerID="6c524eb37d9eb519540115909748ea1e0272427fa38212f1b6de731f3f942edd" Dec 05 10:35:33.320838 master-0 kubenswrapper[4752]: I1205 10:35:33.320552 4752 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 05 10:35:33.334340 master-0 kubenswrapper[4752]: I1205 10:35:33.334306 4752 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 05 10:35:33.641727 master-0 kubenswrapper[4752]: I1205 10:35:33.641594 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:33.753911 master-0 kubenswrapper[4752]: I1205 10:35:33.753845 4752 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:33.758622 master-0 kubenswrapper[4752]: I1205 10:35:33.758534 4752 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:33.871322 master-0 kubenswrapper[4752]: I1205 10:35:33.871259 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerStarted","Data":"303d7cff916f69b65cb8e8b72a47e944bb182f86fc3e9bfde505fa8e541eb3b6"} Dec 05 10:35:33.871569 master-0 kubenswrapper[4752]: I1205 10:35:33.871344 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:33.871569 master-0 kubenswrapper[4752]: I1205 10:35:33.871464 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:33.872242 master-0 kubenswrapper[4752]: I1205 10:35:33.872154 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:33.872242 master-0 kubenswrapper[4752]: I1205 10:35:33.872187 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:33.872242 master-0 kubenswrapper[4752]: I1205 10:35:33.872195 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:34.656932 master-0 kubenswrapper[4752]: I1205 10:35:34.656854 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:34.876448 master-0 kubenswrapper[4752]: I1205 10:35:34.876044 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:34.876448 master-0 kubenswrapper[4752]: I1205 10:35:34.876208 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:34.876448 master-0 kubenswrapper[4752]: I1205 10:35:34.876212 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" event={"ID":"d75143d9bc4a2dc15781dc51ccff632a","Type":"ContainerStarted","Data":"2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef"} Dec 05 10:35:34.877202 master-0 kubenswrapper[4752]: I1205 10:35:34.876924 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:34.877202 master-0 kubenswrapper[4752]: I1205 10:35:34.876954 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:34.877202 master-0 kubenswrapper[4752]: I1205 10:35:34.876964 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:34.877202 master-0 kubenswrapper[4752]: I1205 10:35:34.877076 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:34.877202 master-0 kubenswrapper[4752]: I1205 10:35:34.877116 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:34.877202 master-0 kubenswrapper[4752]: I1205 10:35:34.877129 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:35.371746 master-0 kubenswrapper[4752]: E1205 10:35:35.371539 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51ca172981 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.630762881 +0000 UTC m=+0.170886210,LastTimestamp:2025-12-05 10:35:18.630762881 +0000 UTC m=+0.170886210,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.378335 master-0 kubenswrapper[4752]: E1205 10:35:35.378231 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d025a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node master-0 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68462345 +0000 UTC m=+0.224746769,LastTimestamp:2025-12-05 10:35:18.68462345 +0000 UTC m=+0.224746769,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.383456 master-0 kubenswrapper[4752]: E1205 10:35:35.383346 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d5610 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node master-0 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68464488 +0000 UTC m=+0.224768209,LastTimestamp:2025-12-05 10:35:18.68464488 +0000 UTC m=+0.224768209,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.390021 master-0 kubenswrapper[4752]: E1205 10:35:35.389876 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d749d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node master-0 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.684652701 +0000 UTC m=+0.224776030,LastTimestamp:2025-12-05 10:35:18.684652701 +0000 UTC m=+0.224776030,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.397156 master-0 kubenswrapper[4752]: E1205 10:35:35.396995 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51d2579817 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.769203223 +0000 UTC m=+0.309326552,LastTimestamp:2025-12-05 10:35:18.769203223 +0000 UTC m=+0.309326552,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.403212 master-0 kubenswrapper[4752]: E1205 10:35:35.403119 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d025a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d025a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node master-0 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68462345 +0000 UTC m=+0.224746769,LastTimestamp:2025-12-05 10:35:18.855414725 +0000 UTC m=+0.395538044,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.409993 master-0 kubenswrapper[4752]: E1205 10:35:35.409882 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d5610\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d5610 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node master-0 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68464488 +0000 UTC m=+0.224768209,LastTimestamp:2025-12-05 10:35:18.855451326 +0000 UTC m=+0.395574655,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.418198 master-0 kubenswrapper[4752]: E1205 10:35:35.417930 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d749d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d749d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node master-0 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.684652701 +0000 UTC m=+0.224776030,LastTimestamp:2025-12-05 10:35:18.855459856 +0000 UTC m=+0.395583185,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.426333 master-0 kubenswrapper[4752]: E1205 10:35:35.426108 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d025a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d025a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node master-0 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68462345 +0000 UTC m=+0.224746769,LastTimestamp:2025-12-05 10:35:18.896745468 +0000 UTC m=+0.436868797,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.434434 master-0 kubenswrapper[4752]: E1205 10:35:35.434194 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d5610\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d5610 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node master-0 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68464488 +0000 UTC m=+0.224768209,LastTimestamp:2025-12-05 10:35:18.896765149 +0000 UTC m=+0.436888478,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.441902 master-0 kubenswrapper[4752]: E1205 10:35:35.441758 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d749d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d749d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node master-0 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.684652701 +0000 UTC m=+0.224776030,LastTimestamp:2025-12-05 10:35:18.896773359 +0000 UTC m=+0.436896688,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.447888 master-0 kubenswrapper[4752]: E1205 10:35:35.447727 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d025a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d025a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node master-0 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68462345 +0000 UTC m=+0.224746769,LastTimestamp:2025-12-05 10:35:18.897803355 +0000 UTC m=+0.437926684,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.452008 master-0 kubenswrapper[4752]: E1205 10:35:35.451881 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d5610\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d5610 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node master-0 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68464488 +0000 UTC m=+0.224768209,LastTimestamp:2025-12-05 10:35:18.897824036 +0000 UTC m=+0.437947365,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.457234 master-0 kubenswrapper[4752]: E1205 10:35:35.457063 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d749d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d749d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node master-0 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.684652701 +0000 UTC m=+0.224776030,LastTimestamp:2025-12-05 10:35:18.897837016 +0000 UTC m=+0.437960345,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.461709 master-0 kubenswrapper[4752]: E1205 10:35:35.461581 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d025a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d025a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node master-0 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68462345 +0000 UTC m=+0.224746769,LastTimestamp:2025-12-05 10:35:18.897910449 +0000 UTC m=+0.438033778,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.465641 master-0 kubenswrapper[4752]: E1205 10:35:35.465523 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d5610\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d5610 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node master-0 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68464488 +0000 UTC m=+0.224768209,LastTimestamp:2025-12-05 10:35:18.897926429 +0000 UTC m=+0.438049758,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.469830 master-0 kubenswrapper[4752]: E1205 10:35:35.469706 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d749d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d749d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node master-0 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.684652701 +0000 UTC m=+0.224776030,LastTimestamp:2025-12-05 10:35:18.897937839 +0000 UTC m=+0.438061168,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.473736 master-0 kubenswrapper[4752]: E1205 10:35:35.473636 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d025a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d025a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node master-0 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68462345 +0000 UTC m=+0.224746769,LastTimestamp:2025-12-05 10:35:18.898751978 +0000 UTC m=+0.438875307,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.477446 master-0 kubenswrapper[4752]: E1205 10:35:35.477240 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d5610\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d5610 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node master-0 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68464488 +0000 UTC m=+0.224768209,LastTimestamp:2025-12-05 10:35:18.898768909 +0000 UTC m=+0.438892238,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.481413 master-0 kubenswrapper[4752]: E1205 10:35:35.481308 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d749d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d749d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node master-0 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.684652701 +0000 UTC m=+0.224776030,LastTimestamp:2025-12-05 10:35:18.898776609 +0000 UTC m=+0.438899938,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.485633 master-0 kubenswrapper[4752]: E1205 10:35:35.485296 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d025a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d025a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node master-0 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68462345 +0000 UTC m=+0.224746769,LastTimestamp:2025-12-05 10:35:18.898844701 +0000 UTC m=+0.438968030,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.490186 master-0 kubenswrapper[4752]: E1205 10:35:35.490082 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d5610\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d5610 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node master-0 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68464488 +0000 UTC m=+0.224768209,LastTimestamp:2025-12-05 10:35:18.898854432 +0000 UTC m=+0.438977761,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.494473 master-0 kubenswrapper[4752]: E1205 10:35:35.494304 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d749d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d749d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node master-0 status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.684652701 +0000 UTC m=+0.224776030,LastTimestamp:2025-12-05 10:35:18.898866692 +0000 UTC m=+0.438990021,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.498317 master-0 kubenswrapper[4752]: E1205 10:35:35.498191 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d025a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d025a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node master-0 status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68462345 +0000 UTC m=+0.224746769,LastTimestamp:2025-12-05 10:35:18.899791043 +0000 UTC m=+0.439914372,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.502488 master-0 kubenswrapper[4752]: E1205 10:35:35.502334 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"master-0.187e4b51cd4d5610\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{master-0.187e4b51cd4d5610 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:master-0,UID:master-0,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node master-0 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:18.68464488 +0000 UTC m=+0.224768209,LastTimestamp:2025-12-05 10:35:18.899804874 +0000 UTC m=+0.439928203,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.507783 master-0 kubenswrapper[4752]: E1205 10:35:35.507706 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b5213564125 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulling,Message:Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\",Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:19.859634469 +0000 UTC m=+1.399757798,LastTimestamp:2025-12-05 10:35:19.859634469 +0000 UTC m=+1.399757798,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.511532 master-0 kubenswrapper[4752]: E1205 10:35:35.511447 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{bootstrap-kube-apiserver-master-0.187e4b52146685da openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:bootstrap-kube-apiserver-master-0,UID:d75143d9bc4a2dc15781dc51ccff632a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulling,Message:Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\",Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:19.87747785 +0000 UTC m=+1.417601189,LastTimestamp:2025-12-05 10:35:19.87747785 +0000 UTC m=+1.417601189,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.515741 master-0 kubenswrapper[4752]: E1205 10:35:35.515593 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-scheduler-master-0.187e4b52165b05cf kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-scheduler-master-0,UID:5e09e2af7200e6f9be469dbfd9bb1127,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulling,Message:Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\",Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:19.910278607 +0000 UTC m=+1.450401926,LastTimestamp:2025-12-05 10:35:19.910278607 +0000 UTC m=+1.450401926,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.520526 master-0 kubenswrapper[4752]: E1205 10:35:35.520366 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-controller-manager-master-0.187e4b52172bda2b kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-controller-manager-master-0,UID:8b47694fcc32464ab24d09c23d6efb57,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulling,Message:Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\",Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:19.923964459 +0000 UTC m=+1.464087788,LastTimestamp:2025-12-05 10:35:19.923964459 +0000 UTC m=+1.464087788,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.525274 master-0 kubenswrapper[4752]: E1205 10:35:35.525180 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-master-0-master-0.187e4b52181915cb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-master-0-master-0,UID:cc0396a9a2689b3e8c132c12640cbe83,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulling,Message:Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718\",Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:19.939511755 +0000 UTC m=+1.479635084,LastTimestamp:2025-12-05 10:35:19.939511755 +0000 UTC m=+1.479635084,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.530308 master-0 kubenswrapper[4752]: E1205 10:35:35.530174 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b52c97256a4 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\" in 3.055s (3.055s including waiting). Image size: 459552216 bytes.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:22.914928292 +0000 UTC m=+4.455051621,LastTimestamp:2025-12-05 10:35:22.914928292 +0000 UTC m=+4.455051621,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.535633 master-0 kubenswrapper[4752]: E1205 10:35:35.535540 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-master-0-master-0.187e4b52ca30095a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-master-0-master-0,UID:cc0396a9a2689b3e8c132c12640cbe83,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718\" in 2.987s (2.987s including waiting). Image size: 532719167 bytes.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:22.927360346 +0000 UTC m=+4.467483675,LastTimestamp:2025-12-05 10:35:22.927360346 +0000 UTC m=+4.467483675,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.540624 master-0 kubenswrapper[4752]: E1205 10:35:35.540459 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b52d4fea5c3 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container: setup,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.108672963 +0000 UTC m=+4.648796292,LastTimestamp:2025-12-05 10:35:23.108672963 +0000 UTC m=+4.648796292,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.544776 master-0 kubenswrapper[4752]: E1205 10:35:35.544663 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-master-0-master-0.187e4b52d527d841 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-master-0-master-0,UID:cc0396a9a2689b3e8c132c12640cbe83,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container: etcdctl,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.111372865 +0000 UTC m=+4.651496194,LastTimestamp:2025-12-05 10:35:23.111372865 +0000 UTC m=+4.651496194,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.550242 master-0 kubenswrapper[4752]: E1205 10:35:35.550092 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b52d5d70b91 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.122854801 +0000 UTC m=+4.662978130,LastTimestamp:2025-12-05 10:35:23.122854801 +0000 UTC m=+4.662978130,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.555302 master-0 kubenswrapper[4752]: E1205 10:35:35.555165 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-master-0-master-0.187e4b52d6348248 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-master-0-master-0,UID:cc0396a9a2689b3e8c132c12640cbe83,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.12898004 +0000 UTC m=+4.669103369,LastTimestamp:2025-12-05 10:35:23.12898004 +0000 UTC m=+4.669103369,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.559720 master-0 kubenswrapper[4752]: E1205 10:35:35.559631 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-master-0-master-0.187e4b52d656a8c6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-master-0-master-0,UID:cc0396a9a2689b3e8c132c12640cbe83,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718\" already present on machine,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.131218118 +0000 UTC m=+4.671341457,LastTimestamp:2025-12-05 10:35:23.131218118 +0000 UTC m=+4.671341457,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.564440 master-0 kubenswrapper[4752]: E1205 10:35:35.564316 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-master-0-master-0.187e4b52df85c433 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-master-0-master-0,UID:cc0396a9a2689b3e8c132c12640cbe83,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container: etcd,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.285300275 +0000 UTC m=+4.825423604,LastTimestamp:2025-12-05 10:35:23.285300275 +0000 UTC m=+4.825423604,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.568760 master-0 kubenswrapper[4752]: E1205 10:35:35.568626 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-master-0-master-0.187e4b52e0236d8b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-master-0-master-0,UID:cc0396a9a2689b3e8c132c12640cbe83,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.295632779 +0000 UTC m=+4.835756108,LastTimestamp:2025-12-05 10:35:23.295632779 +0000 UTC m=+4.835756108,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.572555 master-0 kubenswrapper[4752]: E1205 10:35:35.572480 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b52ff2cc4f4 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\" already present on machine,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.816338676 +0000 UTC m=+5.356462005,LastTimestamp:2025-12-05 10:35:23.816338676 +0000 UTC m=+5.356462005,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.578112 master-0 kubenswrapper[4752]: E1205 10:35:35.578008 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b53095de1ce openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container: kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.987329486 +0000 UTC m=+5.527452815,LastTimestamp:2025-12-05 10:35:23.987329486 +0000 UTC m=+5.527452815,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.581695 master-0 kubenswrapper[4752]: E1205 10:35:35.581625 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b530a38fb97 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:24.001688471 +0000 UTC m=+5.541811800,LastTimestamp:2025-12-05 10:35:24.001688471 +0000 UTC m=+5.541811800,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.586038 master-0 kubenswrapper[4752]: E1205 10:35:35.585868 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-rbac-proxy-crio-master-0.187e4b52ff2cc4f4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b52ff2cc4f4 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\" already present on machine,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.816338676 +0000 UTC m=+5.356462005,LastTimestamp:2025-12-05 10:35:28.840744188 +0000 UTC m=+10.380867517,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.590339 master-0 kubenswrapper[4752]: E1205 10:35:35.590227 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-controller-manager-master-0.187e4b542fc5745b kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-controller-manager-master-0,UID:8b47694fcc32464ab24d09c23d6efb57,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\" in 9.002s (9.002s including waiting). Image size: 938303566 bytes.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:28.926618715 +0000 UTC m=+10.466742054,LastTimestamp:2025-12-05 10:35:28.926618715 +0000 UTC m=+10.466742054,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.593832 master-0 kubenswrapper[4752]: E1205 10:35:35.593731 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{bootstrap-kube-apiserver-master-0.187e4b543177dfd3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:bootstrap-kube-apiserver-master-0,UID:d75143d9bc4a2dc15781dc51ccff632a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\" in 9.077s (9.077s including waiting). Image size: 938303566 bytes.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:28.955088851 +0000 UTC m=+10.495212230,LastTimestamp:2025-12-05 10:35:28.955088851 +0000 UTC m=+10.495212230,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.600672 master-0 kubenswrapper[4752]: E1205 10:35:35.600548 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-scheduler-master-0.187e4b5432530f7c kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-scheduler-master-0,UID:5e09e2af7200e6f9be469dbfd9bb1127,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\" in 9.059s (9.059s including waiting). Image size: 938303566 bytes.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:28.969453436 +0000 UTC m=+10.509576805,LastTimestamp:2025-12-05 10:35:28.969453436 +0000 UTC m=+10.509576805,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.604639 master-0 kubenswrapper[4752]: E1205 10:35:35.604515 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-rbac-proxy-crio-master-0.187e4b53095de1ce\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b53095de1ce openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container: kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.987329486 +0000 UTC m=+5.527452815,LastTimestamp:2025-12-05 10:35:29.04318154 +0000 UTC m=+10.583304859,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.608575 master-0 kubenswrapper[4752]: E1205 10:35:35.608475 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-rbac-proxy-crio-master-0.187e4b530a38fb97\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b530a38fb97 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:24.001688471 +0000 UTC m=+5.541811800,LastTimestamp:2025-12-05 10:35:29.056068573 +0000 UTC m=+10.596191902,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.612313 master-0 kubenswrapper[4752]: E1205 10:35:35.612212 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-controller-manager-master-0.187e4b543a8c9bf1 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-controller-manager-master-0,UID:8b47694fcc32464ab24d09c23d6efb57,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container: kube-controller-manager,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.107442673 +0000 UTC m=+10.647565992,LastTimestamp:2025-12-05 10:35:29.107442673 +0000 UTC m=+10.647565992,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.616340 master-0 kubenswrapper[4752]: E1205 10:35:35.616218 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-controller-manager-master-0.187e4b543b26d136 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-controller-manager-master-0,UID:8b47694fcc32464ab24d09c23d6efb57,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.117548854 +0000 UTC m=+10.657672183,LastTimestamp:2025-12-05 10:35:29.117548854 +0000 UTC m=+10.657672183,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.619998 master-0 kubenswrapper[4752]: E1205 10:35:35.619921 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-controller-manager-master-0.187e4b543b3a3081 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-controller-manager-master-0,UID:8b47694fcc32464ab24d09c23d6efb57,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulling,Message:Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96\",Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.118818433 +0000 UTC m=+10.658941762,LastTimestamp:2025-12-05 10:35:29.118818433 +0000 UTC m=+10.658941762,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.624664 master-0 kubenswrapper[4752]: E1205 10:35:35.624512 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-scheduler-master-0.187e4b544074fa3f kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-scheduler-master-0,UID:5e09e2af7200e6f9be469dbfd9bb1127,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container: kube-scheduler,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.206557247 +0000 UTC m=+10.746680576,LastTimestamp:2025-12-05 10:35:29.206557247 +0000 UTC m=+10.746680576,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.628485 master-0 kubenswrapper[4752]: E1205 10:35:35.628389 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{bootstrap-kube-apiserver-master-0.187e4b544083fef3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:bootstrap-kube-apiserver-master-0,UID:d75143d9bc4a2dc15781dc51ccff632a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container: setup,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.207541491 +0000 UTC m=+10.747664840,LastTimestamp:2025-12-05 10:35:29.207541491 +0000 UTC m=+10.747664840,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.632655 master-0 kubenswrapper[4752]: E1205 10:35:35.632568 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-scheduler-master-0.187e4b5441014712 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-scheduler-master-0,UID:5e09e2af7200e6f9be469dbfd9bb1127,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.215751954 +0000 UTC m=+10.755875293,LastTimestamp:2025-12-05 10:35:29.215751954 +0000 UTC m=+10.755875293,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.637516 master-0 kubenswrapper[4752]: I1205 10:35:35.637476 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:35.637691 master-0 kubenswrapper[4752]: E1205 10:35:35.637587 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{bootstrap-kube-apiserver-master-0.187e4b5441463e9c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:bootstrap-kube-apiserver-master-0,UID:d75143d9bc4a2dc15781dc51ccff632a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.220271772 +0000 UTC m=+10.760395111,LastTimestamp:2025-12-05 10:35:29.220271772 +0000 UTC m=+10.760395111,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.642097 master-0 kubenswrapper[4752]: E1205 10:35:35.641992 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{bootstrap-kube-apiserver-master-0.187e4b5466fc6e57 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:bootstrap-kube-apiserver-master-0,UID:d75143d9bc4a2dc15781dc51ccff632a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\" already present on machine,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.852968535 +0000 UTC m=+11.393091864,LastTimestamp:2025-12-05 10:35:29.852968535 +0000 UTC m=+11.393091864,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.646924 master-0 kubenswrapper[4752]: E1205 10:35:35.646782 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b54671784be openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:BackOff,Message:Back-off restarting failed container kube-rbac-proxy-crio in pod kube-rbac-proxy-crio-master-0_openshift-machine-config-operator(3169f44496ed8a28c6d6a15511ab0eec),Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.854743742 +0000 UTC m=+11.394867081,LastTimestamp:2025-12-05 10:35:29.854743742 +0000 UTC m=+11.394867081,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.650915 master-0 kubenswrapper[4752]: E1205 10:35:35.650770 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{bootstrap-kube-apiserver-master-0.187e4b549fa8cfd0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:bootstrap-kube-apiserver-master-0,UID:d75143d9bc4a2dc15781dc51ccff632a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container: kube-apiserver,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:30.803789776 +0000 UTC m=+12.343913105,LastTimestamp:2025-12-05 10:35:30.803789776 +0000 UTC m=+12.343913105,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.655287 master-0 kubenswrapper[4752]: E1205 10:35:35.655160 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-rbac-proxy-crio-master-0.187e4b54671784be\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b54671784be openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:BackOff,Message:Back-off restarting failed container kube-rbac-proxy-crio in pod kube-rbac-proxy-crio-master-0_openshift-machine-config-operator(3169f44496ed8a28c6d6a15511ab0eec),Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.854743742 +0000 UTC m=+11.394867081,LastTimestamp:2025-12-05 10:35:30.858858922 +0000 UTC m=+12.398982251,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.659655 master-0 kubenswrapper[4752]: E1205 10:35:35.659556 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{bootstrap-kube-apiserver-master-0.187e4b54cb982a4b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:bootstrap-kube-apiserver-master-0,UID:d75143d9bc4a2dc15781dc51ccff632a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:31.540896331 +0000 UTC m=+13.081019680,LastTimestamp:2025-12-05 10:35:31.540896331 +0000 UTC m=+13.081019680,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.664217 master-0 kubenswrapper[4752]: E1205 10:35:35.664102 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{bootstrap-kube-apiserver-master-0.187e4b54cba54d8c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:bootstrap-kube-apiserver-master-0,UID:d75143d9bc4a2dc15781dc51ccff632a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulling,Message:Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660\",Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:31.541757324 +0000 UTC m=+13.081880673,LastTimestamp:2025-12-05 10:35:31.541757324 +0000 UTC m=+13.081880673,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.668750 master-0 kubenswrapper[4752]: E1205 10:35:35.668603 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-controller-manager-master-0.187e4b55021ae83b kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-controller-manager-master-0,UID:8b47694fcc32464ab24d09c23d6efb57,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96\" in 3.336s (3.336s including waiting). Image size: 499705918 bytes.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:32.455434299 +0000 UTC m=+13.995557618,LastTimestamp:2025-12-05 10:35:32.455434299 +0000 UTC m=+13.995557618,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.673023 master-0 kubenswrapper[4752]: E1205 10:35:35.672851 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-controller-manager-master-0.187e4b550d85a217 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-controller-manager-master-0,UID:8b47694fcc32464ab24d09c23d6efb57,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container: cluster-policy-controller,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:32.646978071 +0000 UTC m=+14.187101400,LastTimestamp:2025-12-05 10:35:32.646978071 +0000 UTC m=+14.187101400,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.677512 master-0 kubenswrapper[4752]: E1205 10:35:35.677399 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-controller-manager-master-0.187e4b550e18405e kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-controller-manager-master-0,UID:8b47694fcc32464ab24d09c23d6efb57,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:32.656586846 +0000 UTC m=+14.196710175,LastTimestamp:2025-12-05 10:35:32.656586846 +0000 UTC m=+14.196710175,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.681463 master-0 kubenswrapper[4752]: E1205 10:35:35.681346 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-controller-manager-master-0.187e4b551add2fe0 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-controller-manager-master-0,UID:8b47694fcc32464ab24d09c23d6efb57,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\" already present on machine,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:32.870819808 +0000 UTC m=+14.410943137,LastTimestamp:2025-12-05 10:35:32.870819808 +0000 UTC m=+14.410943137,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.686218 master-0 kubenswrapper[4752]: E1205 10:35:35.686142 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"bootstrap-kube-controller-manager-master-0.187e4b543a8c9bf1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-controller-manager-master-0.187e4b543a8c9bf1 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-controller-manager-master-0,UID:8b47694fcc32464ab24d09c23d6efb57,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container: kube-controller-manager,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.107442673 +0000 UTC m=+10.647565992,LastTimestamp:2025-12-05 10:35:33.089645216 +0000 UTC m=+14.629768545,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.689435 master-0 kubenswrapper[4752]: E1205 10:35:35.689339 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"bootstrap-kube-controller-manager-master-0.187e4b543b26d136\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"kube-system\"" event="&Event{ObjectMeta:{bootstrap-kube-controller-manager-master-0.187e4b543b26d136 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:bootstrap-kube-controller-manager-master-0,UID:8b47694fcc32464ab24d09c23d6efb57,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.117548854 +0000 UTC m=+10.657672183,LastTimestamp:2025-12-05 10:35:33.099430336 +0000 UTC m=+14.639553665,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.693322 master-0 kubenswrapper[4752]: E1205 10:35:35.693198 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{bootstrap-kube-apiserver-master-0.187e4b5562516bef openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:bootstrap-kube-apiserver-master-0,UID:d75143d9bc4a2dc15781dc51ccff632a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660\" in 2.527s (2.527s including waiting). Image size: 509437356 bytes.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:34.069619695 +0000 UTC m=+15.609743024,LastTimestamp:2025-12-05 10:35:34.069619695 +0000 UTC m=+15.609743024,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.696613 master-0 kubenswrapper[4752]: E1205 10:35:35.696519 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{bootstrap-kube-apiserver-master-0.187e4b556d4d7d4c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:bootstrap-kube-apiserver-master-0,UID:d75143d9bc4a2dc15781dc51ccff632a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container: kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:34.253911372 +0000 UTC m=+15.794034731,LastTimestamp:2025-12-05 10:35:34.253911372 +0000 UTC m=+15.794034731,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.700108 master-0 kubenswrapper[4752]: E1205 10:35:35.700000 4752 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{bootstrap-kube-apiserver-master-0.187e4b556dd6def2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:bootstrap-kube-apiserver-master-0,UID:d75143d9bc4a2dc15781dc51ccff632a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:34.262914802 +0000 UTC m=+15.803038131,LastTimestamp:2025-12-05 10:35:34.262914802 +0000 UTC m=+15.803038131,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:35.878720 master-0 kubenswrapper[4752]: I1205 10:35:35.878548 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:35.879653 master-0 kubenswrapper[4752]: I1205 10:35:35.879535 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:35.879653 master-0 kubenswrapper[4752]: I1205 10:35:35.879557 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:35.879653 master-0 kubenswrapper[4752]: I1205 10:35:35.879565 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:36.205645 master-0 kubenswrapper[4752]: W1205 10:35:36.205404 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Dec 05 10:35:36.205645 master-0 kubenswrapper[4752]: E1205 10:35:36.205527 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Dec 05 10:35:36.572341 master-0 kubenswrapper[4752]: I1205 10:35:36.572244 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:36.572563 master-0 kubenswrapper[4752]: I1205 10:35:36.572480 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:36.573645 master-0 kubenswrapper[4752]: I1205 10:35:36.573605 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:36.573645 master-0 kubenswrapper[4752]: I1205 10:35:36.573643 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:36.573749 master-0 kubenswrapper[4752]: I1205 10:35:36.573655 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:36.643025 master-0 kubenswrapper[4752]: I1205 10:35:36.642911 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:36.846340 master-0 kubenswrapper[4752]: I1205 10:35:36.846136 4752 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:36.852056 master-0 kubenswrapper[4752]: I1205 10:35:36.851985 4752 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:36.880792 master-0 kubenswrapper[4752]: I1205 10:35:36.880726 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:36.880981 master-0 kubenswrapper[4752]: I1205 10:35:36.880939 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:36.882330 master-0 kubenswrapper[4752]: I1205 10:35:36.882256 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:36.882330 master-0 kubenswrapper[4752]: I1205 10:35:36.882327 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:36.882330 master-0 kubenswrapper[4752]: I1205 10:35:36.882342 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:36.887092 master-0 kubenswrapper[4752]: I1205 10:35:36.887038 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:35:37.644055 master-0 kubenswrapper[4752]: I1205 10:35:37.644006 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:37.883734 master-0 kubenswrapper[4752]: I1205 10:35:37.883663 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:37.884939 master-0 kubenswrapper[4752]: I1205 10:35:37.884859 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:37.884939 master-0 kubenswrapper[4752]: I1205 10:35:37.884919 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:37.885127 master-0 kubenswrapper[4752]: I1205 10:35:37.884961 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:38.377264 master-0 kubenswrapper[4752]: W1205 10:35:38.377184 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "master-0" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Dec 05 10:35:38.377264 master-0 kubenswrapper[4752]: E1205 10:35:38.377258 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"master-0\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Dec 05 10:35:38.643116 master-0 kubenswrapper[4752]: I1205 10:35:38.642974 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:38.756083 master-0 kubenswrapper[4752]: E1205 10:35:38.755981 4752 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"master-0\" not found" Dec 05 10:35:38.886007 master-0 kubenswrapper[4752]: I1205 10:35:38.885941 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:38.886957 master-0 kubenswrapper[4752]: I1205 10:35:38.886878 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:38.886957 master-0 kubenswrapper[4752]: I1205 10:35:38.886939 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:38.886957 master-0 kubenswrapper[4752]: I1205 10:35:38.886956 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:39.226463 master-0 kubenswrapper[4752]: I1205 10:35:39.226358 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:39.227528 master-0 kubenswrapper[4752]: I1205 10:35:39.227460 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:39.227528 master-0 kubenswrapper[4752]: I1205 10:35:39.227517 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:39.227528 master-0 kubenswrapper[4752]: I1205 10:35:39.227540 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:39.227848 master-0 kubenswrapper[4752]: I1205 10:35:39.227620 4752 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:35:39.239169 master-0 kubenswrapper[4752]: E1205 10:35:39.239100 4752 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="master-0" Dec 05 10:35:39.239325 master-0 kubenswrapper[4752]: E1205 10:35:39.239285 4752 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"master-0\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Dec 05 10:35:39.644353 master-0 kubenswrapper[4752]: I1205 10:35:39.644289 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:39.910160 master-0 kubenswrapper[4752]: W1205 10:35:39.910000 4752 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Dec 05 10:35:39.910160 master-0 kubenswrapper[4752]: E1205 10:35:39.910089 4752 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Dec 05 10:35:40.612844 master-0 kubenswrapper[4752]: I1205 10:35:40.612746 4752 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:40.613290 master-0 kubenswrapper[4752]: I1205 10:35:40.612931 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:40.614198 master-0 kubenswrapper[4752]: I1205 10:35:40.614141 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:40.614198 master-0 kubenswrapper[4752]: I1205 10:35:40.614197 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:40.614324 master-0 kubenswrapper[4752]: I1205 10:35:40.614210 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:40.616895 master-0 kubenswrapper[4752]: I1205 10:35:40.616826 4752 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:40.642305 master-0 kubenswrapper[4752]: I1205 10:35:40.642229 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:40.890579 master-0 kubenswrapper[4752]: I1205 10:35:40.890523 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:40.891623 master-0 kubenswrapper[4752]: I1205 10:35:40.891586 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:40.891623 master-0 kubenswrapper[4752]: I1205 10:35:40.891627 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:40.891745 master-0 kubenswrapper[4752]: I1205 10:35:40.891638 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:41.641096 master-0 kubenswrapper[4752]: I1205 10:35:41.640985 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:41.769261 master-0 kubenswrapper[4752]: I1205 10:35:41.769182 4752 csr.go:261] certificate signing request csr-ktl4b is approved, waiting to be issued Dec 05 10:35:42.644748 master-0 kubenswrapper[4752]: I1205 10:35:42.644633 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:43.641795 master-0 kubenswrapper[4752]: I1205 10:35:43.641675 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:44.643203 master-0 kubenswrapper[4752]: I1205 10:35:44.643135 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:44.795969 master-0 kubenswrapper[4752]: I1205 10:35:44.795889 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:44.797170 master-0 kubenswrapper[4752]: I1205 10:35:44.797126 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:44.797239 master-0 kubenswrapper[4752]: I1205 10:35:44.797189 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:44.797239 master-0 kubenswrapper[4752]: I1205 10:35:44.797206 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:44.797786 master-0 kubenswrapper[4752]: I1205 10:35:44.797738 4752 scope.go:117] "RemoveContainer" containerID="aff810b496908f638cc4110ee6780958510ed5ee8e621db2ccc0eb661ea287e6" Dec 05 10:35:44.806883 master-0 kubenswrapper[4752]: E1205 10:35:44.806742 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-rbac-proxy-crio-master-0.187e4b52ff2cc4f4\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b52ff2cc4f4 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\" already present on machine,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.816338676 +0000 UTC m=+5.356462005,LastTimestamp:2025-12-05 10:35:44.800077821 +0000 UTC m=+26.340201170,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:45.109748 master-0 kubenswrapper[4752]: E1205 10:35:45.109598 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-rbac-proxy-crio-master-0.187e4b53095de1ce\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b53095de1ce openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container: kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:23.987329486 +0000 UTC m=+5.527452815,LastTimestamp:2025-12-05 10:35:45.104645609 +0000 UTC m=+26.644768948,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:45.125986 master-0 kubenswrapper[4752]: E1205 10:35:45.125877 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-rbac-proxy-crio-master-0.187e4b530a38fb97\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b530a38fb97 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:24.001688471 +0000 UTC m=+5.541811800,LastTimestamp:2025-12-05 10:35:45.118453732 +0000 UTC m=+26.658577081,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:45.644518 master-0 kubenswrapper[4752]: I1205 10:35:45.644475 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:45.903667 master-0 kubenswrapper[4752]: I1205 10:35:45.903546 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_kube-rbac-proxy-crio-master-0_3169f44496ed8a28c6d6a15511ab0eec/kube-rbac-proxy-crio/2.log" Dec 05 10:35:45.904537 master-0 kubenswrapper[4752]: I1205 10:35:45.904486 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_kube-rbac-proxy-crio-master-0_3169f44496ed8a28c6d6a15511ab0eec/kube-rbac-proxy-crio/1.log" Dec 05 10:35:45.905117 master-0 kubenswrapper[4752]: I1205 10:35:45.905027 4752 generic.go:334] "Generic (PLEG): container finished" podID="3169f44496ed8a28c6d6a15511ab0eec" containerID="c4c2663f7fd33f4762e4466ae563543a600762863049b80576e8bb073c24bbd0" exitCode=1 Dec 05 10:35:45.905117 master-0 kubenswrapper[4752]: I1205 10:35:45.905081 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerDied","Data":"c4c2663f7fd33f4762e4466ae563543a600762863049b80576e8bb073c24bbd0"} Dec 05 10:35:45.905342 master-0 kubenswrapper[4752]: I1205 10:35:45.905151 4752 scope.go:117] "RemoveContainer" containerID="aff810b496908f638cc4110ee6780958510ed5ee8e621db2ccc0eb661ea287e6" Dec 05 10:35:45.905472 master-0 kubenswrapper[4752]: I1205 10:35:45.905393 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:45.907754 master-0 kubenswrapper[4752]: I1205 10:35:45.907243 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:45.907754 master-0 kubenswrapper[4752]: I1205 10:35:45.907288 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:45.907754 master-0 kubenswrapper[4752]: I1205 10:35:45.907338 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:45.908026 master-0 kubenswrapper[4752]: I1205 10:35:45.907886 4752 scope.go:117] "RemoveContainer" containerID="c4c2663f7fd33f4762e4466ae563543a600762863049b80576e8bb073c24bbd0" Dec 05 10:35:45.908231 master-0 kubenswrapper[4752]: E1205 10:35:45.908163 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy-crio\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-rbac-proxy-crio pod=kube-rbac-proxy-crio-master-0_openshift-machine-config-operator(3169f44496ed8a28c6d6a15511ab0eec)\"" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" podUID="3169f44496ed8a28c6d6a15511ab0eec" Dec 05 10:35:45.915291 master-0 kubenswrapper[4752]: E1205 10:35:45.915095 4752 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-rbac-proxy-crio-master-0.187e4b54671784be\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-master-0.187e4b54671784be openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-master-0,UID:3169f44496ed8a28c6d6a15511ab0eec,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:BackOff,Message:Back-off restarting failed container kube-rbac-proxy-crio in pod kube-rbac-proxy-crio-master-0_openshift-machine-config-operator(3169f44496ed8a28c6d6a15511ab0eec),Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:35:29.854743742 +0000 UTC m=+11.394867081,LastTimestamp:2025-12-05 10:35:45.90810262 +0000 UTC m=+27.448225989,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:35:46.240357 master-0 kubenswrapper[4752]: I1205 10:35:46.240210 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:46.241974 master-0 kubenswrapper[4752]: I1205 10:35:46.241892 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:46.242134 master-0 kubenswrapper[4752]: I1205 10:35:46.242015 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:46.242134 master-0 kubenswrapper[4752]: I1205 10:35:46.242090 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:46.242313 master-0 kubenswrapper[4752]: I1205 10:35:46.242207 4752 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:35:46.244943 master-0 kubenswrapper[4752]: E1205 10:35:46.244862 4752 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="master-0" Dec 05 10:35:46.245216 master-0 kubenswrapper[4752]: E1205 10:35:46.245175 4752 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"master-0\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Dec 05 10:35:46.579202 master-0 kubenswrapper[4752]: I1205 10:35:46.579136 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:46.579538 master-0 kubenswrapper[4752]: I1205 10:35:46.579372 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:46.581149 master-0 kubenswrapper[4752]: I1205 10:35:46.581087 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:46.581149 master-0 kubenswrapper[4752]: I1205 10:35:46.581148 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:46.581680 master-0 kubenswrapper[4752]: I1205 10:35:46.581167 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:46.583248 master-0 kubenswrapper[4752]: I1205 10:35:46.583189 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:35:46.638437 master-0 kubenswrapper[4752]: I1205 10:35:46.638352 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:46.909676 master-0 kubenswrapper[4752]: I1205 10:35:46.909530 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_kube-rbac-proxy-crio-master-0_3169f44496ed8a28c6d6a15511ab0eec/kube-rbac-proxy-crio/2.log" Dec 05 10:35:46.910528 master-0 kubenswrapper[4752]: I1205 10:35:46.910338 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:46.911780 master-0 kubenswrapper[4752]: I1205 10:35:46.911727 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:46.911887 master-0 kubenswrapper[4752]: I1205 10:35:46.911791 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:46.911887 master-0 kubenswrapper[4752]: I1205 10:35:46.911813 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:47.643737 master-0 kubenswrapper[4752]: I1205 10:35:47.643557 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:48.643192 master-0 kubenswrapper[4752]: I1205 10:35:48.643070 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:48.756956 master-0 kubenswrapper[4752]: E1205 10:35:48.756814 4752 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"master-0\" not found" Dec 05 10:35:49.642903 master-0 kubenswrapper[4752]: I1205 10:35:49.642849 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:50.645659 master-0 kubenswrapper[4752]: I1205 10:35:50.645553 4752 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "master-0" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 05 10:35:50.902952 master-0 kubenswrapper[4752]: I1205 10:35:50.902839 4752 csr.go:257] certificate signing request csr-ktl4b is issued Dec 05 10:35:51.576867 master-0 kubenswrapper[4752]: I1205 10:35:51.576781 4752 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 05 10:35:51.658960 master-0 kubenswrapper[4752]: I1205 10:35:51.658870 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:51.672629 master-0 kubenswrapper[4752]: I1205 10:35:51.672576 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:51.869662 master-0 kubenswrapper[4752]: I1205 10:35:51.869493 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:51.904192 master-0 kubenswrapper[4752]: I1205 10:35:51.904095 4752 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2025-12-06 10:27:25 +0000 UTC, rotation deadline is 2025-12-06 05:59:22.255503227 +0000 UTC Dec 05 10:35:51.904192 master-0 kubenswrapper[4752]: I1205 10:35:51.904152 4752 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 19h23m30.351355178s for next certificate rotation Dec 05 10:35:52.141595 master-0 kubenswrapper[4752]: I1205 10:35:52.141484 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:52.141595 master-0 kubenswrapper[4752]: E1205 10:35:52.141536 4752 csi_plugin.go:305] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "master-0" not found Dec 05 10:35:52.170314 master-0 kubenswrapper[4752]: I1205 10:35:52.170273 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:52.186973 master-0 kubenswrapper[4752]: I1205 10:35:52.186923 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:52.244625 master-0 kubenswrapper[4752]: I1205 10:35:52.244584 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:52.728880 master-0 kubenswrapper[4752]: I1205 10:35:52.728805 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:52.728880 master-0 kubenswrapper[4752]: E1205 10:35:52.728877 4752 csi_plugin.go:305] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "master-0" not found Dec 05 10:35:52.844170 master-0 kubenswrapper[4752]: I1205 10:35:52.844109 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:52.875525 master-0 kubenswrapper[4752]: I1205 10:35:52.875460 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:52.933359 master-0 kubenswrapper[4752]: I1205 10:35:52.933298 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:53.082208 master-0 kubenswrapper[4752]: I1205 10:35:53.082136 4752 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 10:35:53.191842 master-0 kubenswrapper[4752]: I1205 10:35:53.191784 4752 nodeinfomanager.go:401] Failed to publish CSINode: nodes "master-0" not found Dec 05 10:35:53.191842 master-0 kubenswrapper[4752]: E1205 10:35:53.191820 4752 csi_plugin.go:305] Failed to initialize CSINode: error updating CSINode annotation: timed out waiting for the condition; caused by: nodes "master-0" not found Dec 05 10:35:53.245962 master-0 kubenswrapper[4752]: I1205 10:35:53.245836 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:53.247531 master-0 kubenswrapper[4752]: I1205 10:35:53.247472 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:53.247634 master-0 kubenswrapper[4752]: I1205 10:35:53.247562 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:53.247634 master-0 kubenswrapper[4752]: I1205 10:35:53.247591 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:53.247718 master-0 kubenswrapper[4752]: I1205 10:35:53.247673 4752 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:35:53.251154 master-0 kubenswrapper[4752]: E1205 10:35:53.251099 4752 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"master-0\" not found" node="master-0" Dec 05 10:35:53.255813 master-0 kubenswrapper[4752]: I1205 10:35:53.255756 4752 kubelet_node_status.go:79] "Successfully registered node" node="master-0" Dec 05 10:35:53.255908 master-0 kubenswrapper[4752]: E1205 10:35:53.255839 4752 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": node \"master-0\" not found" Dec 05 10:35:53.266500 master-0 kubenswrapper[4752]: E1205 10:35:53.266454 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:53.367187 master-0 kubenswrapper[4752]: E1205 10:35:53.367036 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:53.467997 master-0 kubenswrapper[4752]: E1205 10:35:53.467918 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:53.569218 master-0 kubenswrapper[4752]: E1205 10:35:53.569155 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:53.658905 master-0 kubenswrapper[4752]: I1205 10:35:53.658721 4752 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Dec 05 10:35:53.666894 master-0 kubenswrapper[4752]: I1205 10:35:53.666838 4752 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 05 10:35:53.669360 master-0 kubenswrapper[4752]: E1205 10:35:53.669308 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:53.770536 master-0 kubenswrapper[4752]: E1205 10:35:53.770458 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:53.871602 master-0 kubenswrapper[4752]: E1205 10:35:53.871499 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:53.972257 master-0 kubenswrapper[4752]: E1205 10:35:53.972038 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:54.072758 master-0 kubenswrapper[4752]: E1205 10:35:54.072674 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:54.173079 master-0 kubenswrapper[4752]: E1205 10:35:54.173012 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:54.274118 master-0 kubenswrapper[4752]: E1205 10:35:54.273952 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:54.374995 master-0 kubenswrapper[4752]: E1205 10:35:54.374902 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:54.475909 master-0 kubenswrapper[4752]: E1205 10:35:54.475820 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:54.576605 master-0 kubenswrapper[4752]: E1205 10:35:54.576518 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:54.677550 master-0 kubenswrapper[4752]: E1205 10:35:54.677493 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:54.778510 master-0 kubenswrapper[4752]: E1205 10:35:54.778448 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:54.879657 master-0 kubenswrapper[4752]: E1205 10:35:54.879505 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:54.980010 master-0 kubenswrapper[4752]: E1205 10:35:54.979929 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:55.081154 master-0 kubenswrapper[4752]: E1205 10:35:55.081002 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:55.181327 master-0 kubenswrapper[4752]: E1205 10:35:55.181169 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:55.281598 master-0 kubenswrapper[4752]: E1205 10:35:55.281534 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:55.382589 master-0 kubenswrapper[4752]: E1205 10:35:55.382517 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:55.483773 master-0 kubenswrapper[4752]: E1205 10:35:55.483635 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:55.584592 master-0 kubenswrapper[4752]: E1205 10:35:55.584387 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:55.684931 master-0 kubenswrapper[4752]: E1205 10:35:55.684834 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:55.785538 master-0 kubenswrapper[4752]: E1205 10:35:55.785453 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:55.886474 master-0 kubenswrapper[4752]: E1205 10:35:55.886343 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:55.925570 master-0 kubenswrapper[4752]: I1205 10:35:55.925488 4752 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 10:35:55.987182 master-0 kubenswrapper[4752]: E1205 10:35:55.987080 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:56.087366 master-0 kubenswrapper[4752]: E1205 10:35:56.087200 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:56.188189 master-0 kubenswrapper[4752]: E1205 10:35:56.188105 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:56.289014 master-0 kubenswrapper[4752]: E1205 10:35:56.288915 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:56.390060 master-0 kubenswrapper[4752]: E1205 10:35:56.389856 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:56.490869 master-0 kubenswrapper[4752]: E1205 10:35:56.490780 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:56.591070 master-0 kubenswrapper[4752]: E1205 10:35:56.590948 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:56.692191 master-0 kubenswrapper[4752]: E1205 10:35:56.692051 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:56.793404 master-0 kubenswrapper[4752]: E1205 10:35:56.793119 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:56.894539 master-0 kubenswrapper[4752]: E1205 10:35:56.894458 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:56.994832 master-0 kubenswrapper[4752]: E1205 10:35:56.994615 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:57.095697 master-0 kubenswrapper[4752]: E1205 10:35:57.095595 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:57.196847 master-0 kubenswrapper[4752]: E1205 10:35:57.196744 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:57.297615 master-0 kubenswrapper[4752]: E1205 10:35:57.297530 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:57.398623 master-0 kubenswrapper[4752]: E1205 10:35:57.398527 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:57.499378 master-0 kubenswrapper[4752]: E1205 10:35:57.499289 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:57.600612 master-0 kubenswrapper[4752]: E1205 10:35:57.600437 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:57.700705 master-0 kubenswrapper[4752]: E1205 10:35:57.700610 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:57.801711 master-0 kubenswrapper[4752]: E1205 10:35:57.801593 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:57.902326 master-0 kubenswrapper[4752]: E1205 10:35:57.902176 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:58.002536 master-0 kubenswrapper[4752]: E1205 10:35:58.002385 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:58.103624 master-0 kubenswrapper[4752]: E1205 10:35:58.103530 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:58.204529 master-0 kubenswrapper[4752]: E1205 10:35:58.204381 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:58.305311 master-0 kubenswrapper[4752]: E1205 10:35:58.305180 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:58.406177 master-0 kubenswrapper[4752]: E1205 10:35:58.406101 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:58.506566 master-0 kubenswrapper[4752]: E1205 10:35:58.506233 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:58.606936 master-0 kubenswrapper[4752]: E1205 10:35:58.606851 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:58.707395 master-0 kubenswrapper[4752]: E1205 10:35:58.707330 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:58.757964 master-0 kubenswrapper[4752]: E1205 10:35:58.757700 4752 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"master-0\" not found" Dec 05 10:35:58.795770 master-0 kubenswrapper[4752]: I1205 10:35:58.795681 4752 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:35:58.796870 master-0 kubenswrapper[4752]: I1205 10:35:58.796822 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:35:58.796956 master-0 kubenswrapper[4752]: I1205 10:35:58.796879 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:35:58.796956 master-0 kubenswrapper[4752]: I1205 10:35:58.796905 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:35:58.797563 master-0 kubenswrapper[4752]: I1205 10:35:58.797500 4752 scope.go:117] "RemoveContainer" containerID="c4c2663f7fd33f4762e4466ae563543a600762863049b80576e8bb073c24bbd0" Dec 05 10:35:58.797789 master-0 kubenswrapper[4752]: E1205 10:35:58.797751 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy-crio\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-rbac-proxy-crio pod=kube-rbac-proxy-crio-master-0_openshift-machine-config-operator(3169f44496ed8a28c6d6a15511ab0eec)\"" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" podUID="3169f44496ed8a28c6d6a15511ab0eec" Dec 05 10:35:58.808553 master-0 kubenswrapper[4752]: E1205 10:35:58.808488 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:58.908722 master-0 kubenswrapper[4752]: E1205 10:35:58.908631 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:59.009132 master-0 kubenswrapper[4752]: E1205 10:35:59.008996 4752 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:35:59.093695 master-0 kubenswrapper[4752]: I1205 10:35:59.093627 4752 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 10:35:59.662764 master-0 kubenswrapper[4752]: I1205 10:35:59.662670 4752 apiserver.go:52] "Watching apiserver" Dec 05 10:35:59.665172 master-0 kubenswrapper[4752]: I1205 10:35:59.665088 4752 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 10:35:59.665480 master-0 kubenswrapper[4752]: I1205 10:35:59.665361 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["assisted-installer/assisted-installer-controller-pd4q6","openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m","openshift-network-operator/network-operator-79767b7ff9-t8j2j"] Dec 05 10:35:59.665880 master-0 kubenswrapper[4752]: I1205 10:35:59.665791 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.665880 master-0 kubenswrapper[4752]: I1205 10:35:59.665851 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.665996 master-0 kubenswrapper[4752]: I1205 10:35:59.665931 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:35:59.668361 master-0 kubenswrapper[4752]: I1205 10:35:59.667785 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 10:35:59.668552 master-0 kubenswrapper[4752]: I1205 10:35:59.668131 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"assisted-installer"/"openshift-service-ca.crt" Dec 05 10:35:59.668612 master-0 kubenswrapper[4752]: I1205 10:35:59.668172 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 10:35:59.668612 master-0 kubenswrapper[4752]: I1205 10:35:59.668594 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 10:35:59.668612 master-0 kubenswrapper[4752]: I1205 10:35:59.668600 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 10:35:59.668701 master-0 kubenswrapper[4752]: I1205 10:35:59.668618 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"assisted-installer"/"assisted-installer-controller-config" Dec 05 10:35:59.668701 master-0 kubenswrapper[4752]: I1205 10:35:59.668665 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"assisted-installer"/"kube-root-ca.crt" Dec 05 10:35:59.669119 master-0 kubenswrapper[4752]: I1205 10:35:59.669074 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 10:35:59.669880 master-0 kubenswrapper[4752]: I1205 10:35:59.669672 4752 reflector.go:368] Caches populated for *v1.Secret from object-"assisted-installer"/"assisted-installer-controller-secret" Dec 05 10:35:59.669880 master-0 kubenswrapper[4752]: I1205 10:35:59.669729 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 10:35:59.738459 master-0 kubenswrapper[4752]: I1205 10:35:59.738361 4752 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Dec 05 10:35:59.828159 master-0 kubenswrapper[4752]: I1205 10:35:59.828061 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-ca-bundle\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-ca-bundle\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.828159 master-0 kubenswrapper[4752]: I1205 10:35:59.828154 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wh8s\" (UniqueName: \"kubernetes.io/projected/f1d74f88-1419-431e-80da-26db419f050e-kube-api-access-2wh8s\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.828973 master-0 kubenswrapper[4752]: I1205 10:35:59.828213 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a0235af-2cf2-4ad4-b419-764fb56a0107-metrics-tls\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:35:59.828973 master-0 kubenswrapper[4752]: I1205 10:35:59.828264 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwqkb\" (UniqueName: \"kubernetes.io/projected/1a0235af-2cf2-4ad4-b419-764fb56a0107-kube-api-access-cwqkb\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:35:59.828973 master-0 kubenswrapper[4752]: I1205 10:35:59.828378 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.828973 master-0 kubenswrapper[4752]: I1205 10:35:59.828618 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-service-ca\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.828973 master-0 kubenswrapper[4752]: I1205 10:35:59.828721 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-run-resolv-conf\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-var-run-resolv-conf\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.828973 master-0 kubenswrapper[4752]: I1205 10:35:59.828761 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-resolv-conf\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-resolv-conf\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.828973 master-0 kubenswrapper[4752]: I1205 10:35:59.828838 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/1a0235af-2cf2-4ad4-b419-764fb56a0107-host-etc-kube\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:35:59.828973 master-0 kubenswrapper[4752]: I1205 10:35:59.828906 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-ssl-certs\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.828973 master-0 kubenswrapper[4752]: I1205 10:35:59.828976 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.829404 master-0 kubenswrapper[4752]: I1205 10:35:59.829047 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-kube-api-access\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.829404 master-0 kubenswrapper[4752]: I1205 10:35:59.829229 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sno-bootstrap-files\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-sno-bootstrap-files\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.930590 master-0 kubenswrapper[4752]: I1205 10:35:59.930279 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-ca-bundle\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-ca-bundle\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.930590 master-0 kubenswrapper[4752]: I1205 10:35:59.930375 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wh8s\" (UniqueName: \"kubernetes.io/projected/f1d74f88-1419-431e-80da-26db419f050e-kube-api-access-2wh8s\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.930590 master-0 kubenswrapper[4752]: I1205 10:35:59.930510 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-ca-bundle\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-ca-bundle\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.930955 master-0 kubenswrapper[4752]: I1205 10:35:59.930631 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a0235af-2cf2-4ad4-b419-764fb56a0107-metrics-tls\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:35:59.930955 master-0 kubenswrapper[4752]: I1205 10:35:59.930829 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwqkb\" (UniqueName: \"kubernetes.io/projected/1a0235af-2cf2-4ad4-b419-764fb56a0107-kube-api-access-cwqkb\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:35:59.931082 master-0 kubenswrapper[4752]: I1205 10:35:59.931038 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.931127 master-0 kubenswrapper[4752]: I1205 10:35:59.931088 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-run-resolv-conf\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-var-run-resolv-conf\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.931167 master-0 kubenswrapper[4752]: I1205 10:35:59.931122 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-resolv-conf\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-resolv-conf\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.931167 master-0 kubenswrapper[4752]: I1205 10:35:59.931158 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-service-ca\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.931230 master-0 kubenswrapper[4752]: I1205 10:35:59.931191 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/1a0235af-2cf2-4ad4-b419-764fb56a0107-host-etc-kube\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:35:59.931261 master-0 kubenswrapper[4752]: I1205 10:35:59.931227 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-ssl-certs\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.931297 master-0 kubenswrapper[4752]: I1205 10:35:59.931263 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.931331 master-0 kubenswrapper[4752]: I1205 10:35:59.931300 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-kube-api-access\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.931622 master-0 kubenswrapper[4752]: I1205 10:35:59.931591 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-resolv-conf\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-resolv-conf\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.931791 master-0 kubenswrapper[4752]: I1205 10:35:59.931738 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/1a0235af-2cf2-4ad4-b419-764fb56a0107-host-etc-kube\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:35:59.931862 master-0 kubenswrapper[4752]: I1205 10:35:59.931814 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sno-bootstrap-files\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-sno-bootstrap-files\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.931968 master-0 kubenswrapper[4752]: I1205 10:35:59.931917 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sno-bootstrap-files\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-sno-bootstrap-files\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.932255 master-0 kubenswrapper[4752]: I1205 10:35:59.932169 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-ssl-certs\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.932379 master-0 kubenswrapper[4752]: E1205 10:35:59.932330 4752 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:35:59.932487 master-0 kubenswrapper[4752]: I1205 10:35:59.932328 4752 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 05 10:35:59.932563 master-0 kubenswrapper[4752]: E1205 10:35:59.932529 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:36:00.432491574 +0000 UTC m=+41.972614943 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:35:59.932634 master-0 kubenswrapper[4752]: I1205 10:35:59.932468 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.932739 master-0 kubenswrapper[4752]: I1205 10:35:59.932686 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-run-resolv-conf\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-var-run-resolv-conf\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.934708 master-0 kubenswrapper[4752]: I1205 10:35:59.934616 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-service-ca\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.938799 master-0 kubenswrapper[4752]: I1205 10:35:59.938747 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a0235af-2cf2-4ad4-b419-764fb56a0107-metrics-tls\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:35:59.969066 master-0 kubenswrapper[4752]: I1205 10:35:59.968954 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wh8s\" (UniqueName: \"kubernetes.io/projected/f1d74f88-1419-431e-80da-26db419f050e-kube-api-access-2wh8s\") pod \"assisted-installer-controller-pd4q6\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:35:59.971654 master-0 kubenswrapper[4752]: I1205 10:35:59.971599 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-kube-api-access\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:35:59.980179 master-0 kubenswrapper[4752]: I1205 10:35:59.980111 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwqkb\" (UniqueName: \"kubernetes.io/projected/1a0235af-2cf2-4ad4-b419-764fb56a0107-kube-api-access-cwqkb\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:35:59.997749 master-0 kubenswrapper[4752]: I1205 10:35:59.997677 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:36:00.014980 master-0 kubenswrapper[4752]: I1205 10:36:00.014903 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:36:00.016860 master-0 kubenswrapper[4752]: W1205 10:36:00.016762 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1d74f88_1419_431e_80da_26db419f050e.slice/crio-e72f14b85f51789b3236420aa213917fe54cddf2ceedaab3f6bad48e1ce12513 WatchSource:0}: Error finding container e72f14b85f51789b3236420aa213917fe54cddf2ceedaab3f6bad48e1ce12513: Status 404 returned error can't find the container with id e72f14b85f51789b3236420aa213917fe54cddf2ceedaab3f6bad48e1ce12513 Dec 05 10:36:00.029992 master-0 kubenswrapper[4752]: W1205 10:36:00.029917 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a0235af_2cf2_4ad4_b419_764fb56a0107.slice/crio-ad7983dc1d046a27e0480393c3ce354f96d0968b079cf24fa1ebf196a469d66d WatchSource:0}: Error finding container ad7983dc1d046a27e0480393c3ce354f96d0968b079cf24fa1ebf196a469d66d: Status 404 returned error can't find the container with id ad7983dc1d046a27e0480393c3ce354f96d0968b079cf24fa1ebf196a469d66d Dec 05 10:36:00.434924 master-0 kubenswrapper[4752]: I1205 10:36:00.434823 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:36:00.435175 master-0 kubenswrapper[4752]: E1205 10:36:00.434996 4752 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:00.435175 master-0 kubenswrapper[4752]: E1205 10:36:00.435061 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:36:01.43504568 +0000 UTC m=+42.975169009 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:00.942557 master-0 kubenswrapper[4752]: I1205 10:36:00.942480 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" event={"ID":"1a0235af-2cf2-4ad4-b419-764fb56a0107","Type":"ContainerStarted","Data":"ad7983dc1d046a27e0480393c3ce354f96d0968b079cf24fa1ebf196a469d66d"} Dec 05 10:36:00.944807 master-0 kubenswrapper[4752]: I1205 10:36:00.944744 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="assisted-installer/assisted-installer-controller-pd4q6" event={"ID":"f1d74f88-1419-431e-80da-26db419f050e","Type":"ContainerStarted","Data":"e72f14b85f51789b3236420aa213917fe54cddf2ceedaab3f6bad48e1ce12513"} Dec 05 10:36:01.441828 master-0 kubenswrapper[4752]: I1205 10:36:01.441726 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:36:01.442375 master-0 kubenswrapper[4752]: E1205 10:36:01.441878 4752 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:01.442375 master-0 kubenswrapper[4752]: E1205 10:36:01.441955 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:36:03.441918715 +0000 UTC m=+44.982042044 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:01.610029 master-0 kubenswrapper[4752]: I1205 10:36:01.609979 4752 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 10:36:02.544706 master-0 kubenswrapper[4752]: I1205 10:36:02.544638 4752 csr.go:261] certificate signing request csr-kkrzt is approved, waiting to be issued Dec 05 10:36:02.554766 master-0 kubenswrapper[4752]: I1205 10:36:02.554712 4752 csr.go:257] certificate signing request csr-kkrzt is issued Dec 05 10:36:03.458707 master-0 kubenswrapper[4752]: I1205 10:36:03.458580 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:36:03.459038 master-0 kubenswrapper[4752]: E1205 10:36:03.458753 4752 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:03.459038 master-0 kubenswrapper[4752]: E1205 10:36:03.458856 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:36:07.458820196 +0000 UTC m=+48.998943545 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:03.556503 master-0 kubenswrapper[4752]: I1205 10:36:03.556301 4752 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2025-12-06 10:27:25 +0000 UTC, rotation deadline is 2025-12-06 07:40:19.803624474 +0000 UTC Dec 05 10:36:03.556503 master-0 kubenswrapper[4752]: I1205 10:36:03.556471 4752 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 21h4m16.247157444s for next certificate rotation Dec 05 10:36:04.556862 master-0 kubenswrapper[4752]: I1205 10:36:04.556804 4752 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2025-12-06 10:27:25 +0000 UTC, rotation deadline is 2025-12-06 06:25:28.16672246 +0000 UTC Dec 05 10:36:04.556862 master-0 kubenswrapper[4752]: I1205 10:36:04.556846 4752 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 19h49m23.60987892s for next certificate rotation Dec 05 10:36:06.959829 master-0 kubenswrapper[4752]: I1205 10:36:06.959733 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" event={"ID":"1a0235af-2cf2-4ad4-b419-764fb56a0107","Type":"ContainerStarted","Data":"f1e6d80e2ae7d337901560dde2050e55f54f6b182ab45a5e1cce9ebc5f4043a9"} Dec 05 10:36:06.961230 master-0 kubenswrapper[4752]: I1205 10:36:06.961191 4752 generic.go:334] "Generic (PLEG): container finished" podID="f1d74f88-1419-431e-80da-26db419f050e" containerID="afde21a1bd7da8808580dd974998cee92cb64026aa792ec094def214f272bc2f" exitCode=0 Dec 05 10:36:06.961230 master-0 kubenswrapper[4752]: I1205 10:36:06.961221 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="assisted-installer/assisted-installer-controller-pd4q6" event={"ID":"f1d74f88-1419-431e-80da-26db419f050e","Type":"ContainerDied","Data":"afde21a1bd7da8808580dd974998cee92cb64026aa792ec094def214f272bc2f"} Dec 05 10:36:07.024705 master-0 kubenswrapper[4752]: I1205 10:36:07.024640 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" podStartSLOduration=8.109560506 podStartE2EDuration="14.024624408s" podCreationTimestamp="2025-12-05 10:35:53 +0000 UTC" firstStartedPulling="2025-12-05 10:36:00.032912498 +0000 UTC m=+41.573035827" lastFinishedPulling="2025-12-05 10:36:05.94797638 +0000 UTC m=+47.488099729" observedRunningTime="2025-12-05 10:36:07.024590777 +0000 UTC m=+48.564714126" watchObservedRunningTime="2025-12-05 10:36:07.024624408 +0000 UTC m=+48.564747737" Dec 05 10:36:07.489638 master-0 kubenswrapper[4752]: I1205 10:36:07.489506 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:36:07.490057 master-0 kubenswrapper[4752]: E1205 10:36:07.489731 4752 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:07.490057 master-0 kubenswrapper[4752]: E1205 10:36:07.489888 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:36:15.489864319 +0000 UTC m=+57.029987658 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:07.979173 master-0 kubenswrapper[4752]: I1205 10:36:07.979128 4752 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:36:08.096512 master-0 kubenswrapper[4752]: I1205 10:36:08.096411 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wh8s\" (UniqueName: \"kubernetes.io/projected/f1d74f88-1419-431e-80da-26db419f050e-kube-api-access-2wh8s\") pod \"f1d74f88-1419-431e-80da-26db419f050e\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " Dec 05 10:36:08.096512 master-0 kubenswrapper[4752]: I1205 10:36:08.096503 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-run-resolv-conf\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-var-run-resolv-conf\") pod \"f1d74f88-1419-431e-80da-26db419f050e\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " Dec 05 10:36:08.096512 master-0 kubenswrapper[4752]: I1205 10:36:08.096521 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sno-bootstrap-files\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-sno-bootstrap-files\") pod \"f1d74f88-1419-431e-80da-26db419f050e\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " Dec 05 10:36:08.097223 master-0 kubenswrapper[4752]: I1205 10:36:08.096538 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-ca-bundle\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-ca-bundle\") pod \"f1d74f88-1419-431e-80da-26db419f050e\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " Dec 05 10:36:08.097223 master-0 kubenswrapper[4752]: I1205 10:36:08.096560 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-resolv-conf\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-resolv-conf\") pod \"f1d74f88-1419-431e-80da-26db419f050e\" (UID: \"f1d74f88-1419-431e-80da-26db419f050e\") " Dec 05 10:36:08.097223 master-0 kubenswrapper[4752]: I1205 10:36:08.096636 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-var-run-resolv-conf" (OuterVolumeSpecName: "host-var-run-resolv-conf") pod "f1d74f88-1419-431e-80da-26db419f050e" (UID: "f1d74f88-1419-431e-80da-26db419f050e"). InnerVolumeSpecName "host-var-run-resolv-conf". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:36:08.097223 master-0 kubenswrapper[4752]: I1205 10:36:08.096675 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-sno-bootstrap-files" (OuterVolumeSpecName: "sno-bootstrap-files") pod "f1d74f88-1419-431e-80da-26db419f050e" (UID: "f1d74f88-1419-431e-80da-26db419f050e"). InnerVolumeSpecName "sno-bootstrap-files". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:36:08.097223 master-0 kubenswrapper[4752]: I1205 10:36:08.096707 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-resolv-conf" (OuterVolumeSpecName: "host-resolv-conf") pod "f1d74f88-1419-431e-80da-26db419f050e" (UID: "f1d74f88-1419-431e-80da-26db419f050e"). InnerVolumeSpecName "host-resolv-conf". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:36:08.097223 master-0 kubenswrapper[4752]: I1205 10:36:08.096775 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-ca-bundle" (OuterVolumeSpecName: "host-ca-bundle") pod "f1d74f88-1419-431e-80da-26db419f050e" (UID: "f1d74f88-1419-431e-80da-26db419f050e"). InnerVolumeSpecName "host-ca-bundle". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:36:08.099815 master-0 kubenswrapper[4752]: I1205 10:36:08.099761 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1d74f88-1419-431e-80da-26db419f050e-kube-api-access-2wh8s" (OuterVolumeSpecName: "kube-api-access-2wh8s") pod "f1d74f88-1419-431e-80da-26db419f050e" (UID: "f1d74f88-1419-431e-80da-26db419f050e"). InnerVolumeSpecName "kube-api-access-2wh8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:36:08.197832 master-0 kubenswrapper[4752]: I1205 10:36:08.197724 4752 reconciler_common.go:293] "Volume detached for volume \"host-var-run-resolv-conf\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-var-run-resolv-conf\") on node \"master-0\" DevicePath \"\"" Dec 05 10:36:08.197832 master-0 kubenswrapper[4752]: I1205 10:36:08.197768 4752 reconciler_common.go:293] "Volume detached for volume \"host-ca-bundle\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:36:08.197832 master-0 kubenswrapper[4752]: I1205 10:36:08.197780 4752 reconciler_common.go:293] "Volume detached for volume \"host-resolv-conf\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-host-resolv-conf\") on node \"master-0\" DevicePath \"\"" Dec 05 10:36:08.197832 master-0 kubenswrapper[4752]: I1205 10:36:08.197793 4752 reconciler_common.go:293] "Volume detached for volume \"sno-bootstrap-files\" (UniqueName: \"kubernetes.io/host-path/f1d74f88-1419-431e-80da-26db419f050e-sno-bootstrap-files\") on node \"master-0\" DevicePath \"\"" Dec 05 10:36:08.197832 master-0 kubenswrapper[4752]: I1205 10:36:08.197804 4752 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wh8s\" (UniqueName: \"kubernetes.io/projected/f1d74f88-1419-431e-80da-26db419f050e-kube-api-access-2wh8s\") on node \"master-0\" DevicePath \"\"" Dec 05 10:36:08.966459 master-0 kubenswrapper[4752]: I1205 10:36:08.966347 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="assisted-installer/assisted-installer-controller-pd4q6" event={"ID":"f1d74f88-1419-431e-80da-26db419f050e","Type":"ContainerDied","Data":"e72f14b85f51789b3236420aa213917fe54cddf2ceedaab3f6bad48e1ce12513"} Dec 05 10:36:08.966459 master-0 kubenswrapper[4752]: I1205 10:36:08.966395 4752 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:36:08.966752 master-0 kubenswrapper[4752]: I1205 10:36:08.966407 4752 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e72f14b85f51789b3236420aa213917fe54cddf2ceedaab3f6bad48e1ce12513" Dec 05 10:36:09.332133 master-0 kubenswrapper[4752]: I1205 10:36:09.332085 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/mtu-prober-4w5fd"] Dec 05 10:36:09.332752 master-0 kubenswrapper[4752]: E1205 10:36:09.332171 4752 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1d74f88-1419-431e-80da-26db419f050e" containerName="assisted-installer-controller" Dec 05 10:36:09.332752 master-0 kubenswrapper[4752]: I1205 10:36:09.332185 4752 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1d74f88-1419-431e-80da-26db419f050e" containerName="assisted-installer-controller" Dec 05 10:36:09.332752 master-0 kubenswrapper[4752]: I1205 10:36:09.332211 4752 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1d74f88-1419-431e-80da-26db419f050e" containerName="assisted-installer-controller" Dec 05 10:36:09.332752 master-0 kubenswrapper[4752]: I1205 10:36:09.332387 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/mtu-prober-4w5fd" Dec 05 10:36:09.509050 master-0 kubenswrapper[4752]: I1205 10:36:09.508904 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6djnn\" (UniqueName: \"kubernetes.io/projected/3af980e3-b7b6-4a5b-96c4-53cbbe43c268-kube-api-access-6djnn\") pod \"mtu-prober-4w5fd\" (UID: \"3af980e3-b7b6-4a5b-96c4-53cbbe43c268\") " pod="openshift-network-operator/mtu-prober-4w5fd" Dec 05 10:36:09.610473 master-0 kubenswrapper[4752]: I1205 10:36:09.610249 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6djnn\" (UniqueName: \"kubernetes.io/projected/3af980e3-b7b6-4a5b-96c4-53cbbe43c268-kube-api-access-6djnn\") pod \"mtu-prober-4w5fd\" (UID: \"3af980e3-b7b6-4a5b-96c4-53cbbe43c268\") " pod="openshift-network-operator/mtu-prober-4w5fd" Dec 05 10:36:09.636859 master-0 kubenswrapper[4752]: I1205 10:36:09.636781 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6djnn\" (UniqueName: \"kubernetes.io/projected/3af980e3-b7b6-4a5b-96c4-53cbbe43c268-kube-api-access-6djnn\") pod \"mtu-prober-4w5fd\" (UID: \"3af980e3-b7b6-4a5b-96c4-53cbbe43c268\") " pod="openshift-network-operator/mtu-prober-4w5fd" Dec 05 10:36:09.643382 master-0 kubenswrapper[4752]: I1205 10:36:09.643345 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/mtu-prober-4w5fd" Dec 05 10:36:09.654831 master-0 kubenswrapper[4752]: W1205 10:36:09.654741 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3af980e3_b7b6_4a5b_96c4_53cbbe43c268.slice/crio-2323b8713dbbb688b1af973566334ca0b77411424d2f12acb62f9e5354b6cd41 WatchSource:0}: Error finding container 2323b8713dbbb688b1af973566334ca0b77411424d2f12acb62f9e5354b6cd41: Status 404 returned error can't find the container with id 2323b8713dbbb688b1af973566334ca0b77411424d2f12acb62f9e5354b6cd41 Dec 05 10:36:09.970016 master-0 kubenswrapper[4752]: I1205 10:36:09.969776 4752 generic.go:334] "Generic (PLEG): container finished" podID="3af980e3-b7b6-4a5b-96c4-53cbbe43c268" containerID="53136dd9eb8d1d61f6ae883e0b488e0fb1c3b2414f554f6f62aa4dc092f5012c" exitCode=0 Dec 05 10:36:09.970016 master-0 kubenswrapper[4752]: I1205 10:36:09.969993 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/mtu-prober-4w5fd" event={"ID":"3af980e3-b7b6-4a5b-96c4-53cbbe43c268","Type":"ContainerDied","Data":"53136dd9eb8d1d61f6ae883e0b488e0fb1c3b2414f554f6f62aa4dc092f5012c"} Dec 05 10:36:09.970143 master-0 kubenswrapper[4752]: I1205 10:36:09.970037 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/mtu-prober-4w5fd" event={"ID":"3af980e3-b7b6-4a5b-96c4-53cbbe43c268","Type":"ContainerStarted","Data":"2323b8713dbbb688b1af973566334ca0b77411424d2f12acb62f9e5354b6cd41"} Dec 05 10:36:10.986824 master-0 kubenswrapper[4752]: I1205 10:36:10.986784 4752 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/mtu-prober-4w5fd" Dec 05 10:36:11.123170 master-0 kubenswrapper[4752]: I1205 10:36:11.123069 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6djnn\" (UniqueName: \"kubernetes.io/projected/3af980e3-b7b6-4a5b-96c4-53cbbe43c268-kube-api-access-6djnn\") pod \"3af980e3-b7b6-4a5b-96c4-53cbbe43c268\" (UID: \"3af980e3-b7b6-4a5b-96c4-53cbbe43c268\") " Dec 05 10:36:11.128077 master-0 kubenswrapper[4752]: I1205 10:36:11.128012 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3af980e3-b7b6-4a5b-96c4-53cbbe43c268-kube-api-access-6djnn" (OuterVolumeSpecName: "kube-api-access-6djnn") pod "3af980e3-b7b6-4a5b-96c4-53cbbe43c268" (UID: "3af980e3-b7b6-4a5b-96c4-53cbbe43c268"). InnerVolumeSpecName "kube-api-access-6djnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:36:11.223833 master-0 kubenswrapper[4752]: I1205 10:36:11.223601 4752 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6djnn\" (UniqueName: \"kubernetes.io/projected/3af980e3-b7b6-4a5b-96c4-53cbbe43c268-kube-api-access-6djnn\") on node \"master-0\" DevicePath \"\"" Dec 05 10:36:11.810077 master-0 kubenswrapper[4752]: I1205 10:36:11.809984 4752 scope.go:117] "RemoveContainer" containerID="c4c2663f7fd33f4762e4466ae563543a600762863049b80576e8bb073c24bbd0" Dec 05 10:36:11.810283 master-0 kubenswrapper[4752]: I1205 10:36:11.809983 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-master-0"] Dec 05 10:36:11.976636 master-0 kubenswrapper[4752]: I1205 10:36:11.976600 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/mtu-prober-4w5fd" event={"ID":"3af980e3-b7b6-4a5b-96c4-53cbbe43c268","Type":"ContainerDied","Data":"2323b8713dbbb688b1af973566334ca0b77411424d2f12acb62f9e5354b6cd41"} Dec 05 10:36:11.977349 master-0 kubenswrapper[4752]: I1205 10:36:11.977213 4752 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2323b8713dbbb688b1af973566334ca0b77411424d2f12acb62f9e5354b6cd41" Dec 05 10:36:11.977349 master-0 kubenswrapper[4752]: I1205 10:36:11.976633 4752 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/mtu-prober-4w5fd" Dec 05 10:36:12.981655 master-0 kubenswrapper[4752]: I1205 10:36:12.981611 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_kube-rbac-proxy-crio-master-0_3169f44496ed8a28c6d6a15511ab0eec/kube-rbac-proxy-crio/2.log" Dec 05 10:36:12.982126 master-0 kubenswrapper[4752]: I1205 10:36:12.982032 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerStarted","Data":"3a88ca4d52cee88aa8dcb914ea7d8cff8936f823056376f936c3da2aa8e9a481"} Dec 05 10:36:13.074734 master-0 kubenswrapper[4752]: I1205 10:36:13.074647 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" podStartSLOduration=2.074625755 podStartE2EDuration="2.074625755s" podCreationTimestamp="2025-12-05 10:36:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:36:13.074332978 +0000 UTC m=+54.614456327" watchObservedRunningTime="2025-12-05 10:36:13.074625755 +0000 UTC m=+54.614749084" Dec 05 10:36:14.324907 master-0 kubenswrapper[4752]: I1205 10:36:14.324864 4752 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-network-operator/mtu-prober-4w5fd"] Dec 05 10:36:14.327462 master-0 kubenswrapper[4752]: I1205 10:36:14.327257 4752 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-network-operator/mtu-prober-4w5fd"] Dec 05 10:36:14.800274 master-0 kubenswrapper[4752]: I1205 10:36:14.800199 4752 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3af980e3-b7b6-4a5b-96c4-53cbbe43c268" path="/var/lib/kubelet/pods/3af980e3-b7b6-4a5b-96c4-53cbbe43c268/volumes" Dec 05 10:36:15.553236 master-0 kubenswrapper[4752]: I1205 10:36:15.553151 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:36:15.553809 master-0 kubenswrapper[4752]: E1205 10:36:15.553394 4752 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:15.553809 master-0 kubenswrapper[4752]: E1205 10:36:15.553537 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:36:31.553515208 +0000 UTC m=+73.093638537 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:19.185157 master-0 kubenswrapper[4752]: I1205 10:36:19.184962 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-lxmgz"] Dec 05 10:36:19.185157 master-0 kubenswrapper[4752]: E1205 10:36:19.185051 4752 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af980e3-b7b6-4a5b-96c4-53cbbe43c268" containerName="prober" Dec 05 10:36:19.185157 master-0 kubenswrapper[4752]: I1205 10:36:19.185065 4752 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af980e3-b7b6-4a5b-96c4-53cbbe43c268" containerName="prober" Dec 05 10:36:19.185157 master-0 kubenswrapper[4752]: I1205 10:36:19.185093 4752 memory_manager.go:354] "RemoveStaleState removing state" podUID="3af980e3-b7b6-4a5b-96c4-53cbbe43c268" containerName="prober" Dec 05 10:36:19.186049 master-0 kubenswrapper[4752]: I1205 10:36:19.185281 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.187112 master-0 kubenswrapper[4752]: I1205 10:36:19.187078 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 10:36:19.187599 master-0 kubenswrapper[4752]: I1205 10:36:19.187558 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 10:36:19.188950 master-0 kubenswrapper[4752]: I1205 10:36:19.188898 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 10:36:19.189082 master-0 kubenswrapper[4752]: I1205 10:36:19.189051 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 10:36:19.378943 master-0 kubenswrapper[4752]: I1205 10:36:19.378823 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-multus\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.378943 master-0 kubenswrapper[4752]: I1205 10:36:19.378915 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-kubelet\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379352 master-0 kubenswrapper[4752]: I1205 10:36:19.378975 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-multus-daemon-config\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379352 master-0 kubenswrapper[4752]: I1205 10:36:19.379021 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-etc-kubernetes\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379352 master-0 kubenswrapper[4752]: I1205 10:36:19.379064 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-socket-dir-parent\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379352 master-0 kubenswrapper[4752]: I1205 10:36:19.379110 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-bin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379352 master-0 kubenswrapper[4752]: I1205 10:36:19.379184 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-os-release\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379352 master-0 kubenswrapper[4752]: I1205 10:36:19.379224 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379352 master-0 kubenswrapper[4752]: I1205 10:36:19.379332 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-k8s-cni-cncf-io\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379940 master-0 kubenswrapper[4752]: I1205 10:36:19.379405 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-netns\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379940 master-0 kubenswrapper[4752]: I1205 10:36:19.379502 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-conf-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379940 master-0 kubenswrapper[4752]: I1205 10:36:19.379550 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-cni-binary-copy\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379940 master-0 kubenswrapper[4752]: I1205 10:36:19.379594 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-hostroot\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379940 master-0 kubenswrapper[4752]: I1205 10:36:19.379636 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvnxf\" (UniqueName: \"kubernetes.io/projected/6a9f011f-36f1-4308-a365-69425c186c7f-kube-api-access-mvnxf\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379940 master-0 kubenswrapper[4752]: I1205 10:36:19.379683 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-system-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379940 master-0 kubenswrapper[4752]: I1205 10:36:19.379725 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-cnibin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.379940 master-0 kubenswrapper[4752]: I1205 10:36:19.379848 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-multus-certs\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.389266 master-0 kubenswrapper[4752]: I1205 10:36:19.389220 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-dms5d"] Dec 05 10:36:19.389892 master-0 kubenswrapper[4752]: I1205 10:36:19.389838 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.392215 master-0 kubenswrapper[4752]: I1205 10:36:19.392157 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"whereabouts-config" Dec 05 10:36:19.392659 master-0 kubenswrapper[4752]: I1205 10:36:19.392627 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480211 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-cnibin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480296 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-system-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480332 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-multus-certs\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480351 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-cnibin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480365 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-multus\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480453 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-multus\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480453 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-kubelet\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480483 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-kubelet\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480501 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-multus-daemon-config\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480533 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-bin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480563 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-etc-kubernetes\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480573 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-multus-certs\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480586 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-socket-dir-parent\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480648 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-socket-dir-parent\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480649 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-os-release\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480681 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480705 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-k8s-cni-cncf-io\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.480905 master-0 kubenswrapper[4752]: I1205 10:36:19.480711 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-os-release\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.480730 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-netns\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.480947 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-conf-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.480945 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-system-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.480983 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-cni-binary-copy\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.481119 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-hostroot\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.481156 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvnxf\" (UniqueName: \"kubernetes.io/projected/6a9f011f-36f1-4308-a365-69425c186c7f-kube-api-access-mvnxf\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.481651 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-etc-kubernetes\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.481705 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-conf-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.481761 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-bin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.481801 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-hostroot\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.481838 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-k8s-cni-cncf-io\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.481838 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482145 master-0 kubenswrapper[4752]: I1205 10:36:19.481894 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-cni-binary-copy\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482990 master-0 kubenswrapper[4752]: I1205 10:36:19.482563 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-multus-daemon-config\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.482990 master-0 kubenswrapper[4752]: I1205 10:36:19.482615 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-netns\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.506566 master-0 kubenswrapper[4752]: I1205 10:36:19.506483 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvnxf\" (UniqueName: \"kubernetes.io/projected/6a9f011f-36f1-4308-a365-69425c186c7f-kube-api-access-mvnxf\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.581542 master-0 kubenswrapper[4752]: I1205 10:36:19.581459 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cnibin\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.581542 master-0 kubenswrapper[4752]: I1205 10:36:19.581531 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-binary-copy\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.581916 master-0 kubenswrapper[4752]: I1205 10:36:19.581559 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.581916 master-0 kubenswrapper[4752]: I1205 10:36:19.581786 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-configmap\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-whereabouts-configmap\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.582066 master-0 kubenswrapper[4752]: I1205 10:36:19.581916 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrdsv\" (UniqueName: \"kubernetes.io/projected/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-kube-api-access-jrdsv\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.582066 master-0 kubenswrapper[4752]: I1205 10:36:19.581949 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-os-release\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.582066 master-0 kubenswrapper[4752]: I1205 10:36:19.581973 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-system-cni-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.582066 master-0 kubenswrapper[4752]: I1205 10:36:19.582030 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.683219 master-0 kubenswrapper[4752]: I1205 10:36:19.683105 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-system-cni-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.683219 master-0 kubenswrapper[4752]: I1205 10:36:19.683171 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.683219 master-0 kubenswrapper[4752]: I1205 10:36:19.683206 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cnibin\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.683219 master-0 kubenswrapper[4752]: I1205 10:36:19.683228 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-binary-copy\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.683920 master-0 kubenswrapper[4752]: I1205 10:36:19.683251 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.683920 master-0 kubenswrapper[4752]: I1205 10:36:19.683373 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-system-cni-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.683920 master-0 kubenswrapper[4752]: I1205 10:36:19.683640 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"whereabouts-configmap\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-whereabouts-configmap\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.683920 master-0 kubenswrapper[4752]: I1205 10:36:19.683722 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrdsv\" (UniqueName: \"kubernetes.io/projected/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-kube-api-access-jrdsv\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.683920 master-0 kubenswrapper[4752]: I1205 10:36:19.683745 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-os-release\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.683920 master-0 kubenswrapper[4752]: I1205 10:36:19.683864 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-os-release\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.683920 master-0 kubenswrapper[4752]: I1205 10:36:19.683882 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.684329 master-0 kubenswrapper[4752]: I1205 10:36:19.684107 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cnibin\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.684549 master-0 kubenswrapper[4752]: I1205 10:36:19.684503 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"whereabouts-configmap\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-whereabouts-configmap\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.684971 master-0 kubenswrapper[4752]: I1205 10:36:19.684575 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.685064 master-0 kubenswrapper[4752]: I1205 10:36:19.685007 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-binary-copy\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.700616 master-0 kubenswrapper[4752]: I1205 10:36:19.700557 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrdsv\" (UniqueName: \"kubernetes.io/projected/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-kube-api-access-jrdsv\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.710475 master-0 kubenswrapper[4752]: I1205 10:36:19.710367 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:36:19.729240 master-0 kubenswrapper[4752]: W1205 10:36:19.729141 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38ad6e6a_a2b9_44e9_ac0b_e413c65efad8.slice/crio-efc2234f689663f4f73596bfe5f7a66235fb095a2337fa0c8bc412e1e08433fb WatchSource:0}: Error finding container efc2234f689663f4f73596bfe5f7a66235fb095a2337fa0c8bc412e1e08433fb: Status 404 returned error can't find the container with id efc2234f689663f4f73596bfe5f7a66235fb095a2337fa0c8bc412e1e08433fb Dec 05 10:36:19.799084 master-0 kubenswrapper[4752]: I1205 10:36:19.799020 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-lxmgz" Dec 05 10:36:19.812635 master-0 kubenswrapper[4752]: W1205 10:36:19.812584 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a9f011f_36f1_4308_a365_69425c186c7f.slice/crio-8d62f303e35a44bd684cd9b7f51835e4e80959496c433fcc66b39cdd84c84cc8 WatchSource:0}: Error finding container 8d62f303e35a44bd684cd9b7f51835e4e80959496c433fcc66b39cdd84c84cc8: Status 404 returned error can't find the container with id 8d62f303e35a44bd684cd9b7f51835e4e80959496c433fcc66b39cdd84c84cc8 Dec 05 10:36:19.999058 master-0 kubenswrapper[4752]: I1205 10:36:19.998911 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dms5d" event={"ID":"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8","Type":"ContainerStarted","Data":"efc2234f689663f4f73596bfe5f7a66235fb095a2337fa0c8bc412e1e08433fb"} Dec 05 10:36:19.999656 master-0 kubenswrapper[4752]: I1205 10:36:19.999629 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lxmgz" event={"ID":"6a9f011f-36f1-4308-a365-69425c186c7f","Type":"ContainerStarted","Data":"8d62f303e35a44bd684cd9b7f51835e4e80959496c433fcc66b39cdd84c84cc8"} Dec 05 10:36:20.181484 master-0 kubenswrapper[4752]: I1205 10:36:20.181352 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-8gjgm"] Dec 05 10:36:20.182388 master-0 kubenswrapper[4752]: I1205 10:36:20.182335 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:20.182510 master-0 kubenswrapper[4752]: E1205 10:36:20.182455 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:20.289147 master-0 kubenswrapper[4752]: I1205 10:36:20.289047 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:20.289806 master-0 kubenswrapper[4752]: I1205 10:36:20.289160 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssssf\" (UniqueName: \"kubernetes.io/projected/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-kube-api-access-ssssf\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:20.390363 master-0 kubenswrapper[4752]: I1205 10:36:20.390299 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssssf\" (UniqueName: \"kubernetes.io/projected/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-kube-api-access-ssssf\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:20.390363 master-0 kubenswrapper[4752]: I1205 10:36:20.390360 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:20.390696 master-0 kubenswrapper[4752]: E1205 10:36:20.390665 4752 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:20.390747 master-0 kubenswrapper[4752]: E1205 10:36:20.390738 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:36:20.8907187 +0000 UTC m=+62.430842039 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:20.411791 master-0 kubenswrapper[4752]: I1205 10:36:20.411740 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssssf\" (UniqueName: \"kubernetes.io/projected/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-kube-api-access-ssssf\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:20.895460 master-0 kubenswrapper[4752]: I1205 10:36:20.895382 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:20.895667 master-0 kubenswrapper[4752]: E1205 10:36:20.895578 4752 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:20.895667 master-0 kubenswrapper[4752]: E1205 10:36:20.895647 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:36:21.895627746 +0000 UTC m=+63.435751065 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:21.795001 master-0 kubenswrapper[4752]: I1205 10:36:21.794925 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:21.795643 master-0 kubenswrapper[4752]: E1205 10:36:21.795094 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:21.903712 master-0 kubenswrapper[4752]: I1205 10:36:21.903587 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:21.903957 master-0 kubenswrapper[4752]: E1205 10:36:21.903794 4752 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:21.903957 master-0 kubenswrapper[4752]: E1205 10:36:21.903856 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:36:23.903835045 +0000 UTC m=+65.443958374 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:23.796019 master-0 kubenswrapper[4752]: I1205 10:36:23.795736 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:23.796511 master-0 kubenswrapper[4752]: E1205 10:36:23.796075 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:23.919128 master-0 kubenswrapper[4752]: I1205 10:36:23.919054 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:23.919340 master-0 kubenswrapper[4752]: E1205 10:36:23.919218 4752 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:23.919340 master-0 kubenswrapper[4752]: E1205 10:36:23.919286 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:36:27.919268349 +0000 UTC m=+69.459391678 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:24.013132 master-0 kubenswrapper[4752]: I1205 10:36:24.013069 4752 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="10b806d181204c81aa236ce7e2f93aa28932b9c9f02c11fefdb2f0d1d7bdeac9" exitCode=0 Dec 05 10:36:24.013132 master-0 kubenswrapper[4752]: I1205 10:36:24.013116 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dms5d" event={"ID":"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8","Type":"ContainerDied","Data":"10b806d181204c81aa236ce7e2f93aa28932b9c9f02c11fefdb2f0d1d7bdeac9"} Dec 05 10:36:25.796173 master-0 kubenswrapper[4752]: I1205 10:36:25.796000 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:25.798323 master-0 kubenswrapper[4752]: E1205 10:36:25.796338 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:27.795750 master-0 kubenswrapper[4752]: I1205 10:36:27.795643 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:27.796940 master-0 kubenswrapper[4752]: E1205 10:36:27.795812 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:27.955946 master-0 kubenswrapper[4752]: I1205 10:36:27.955892 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:27.956129 master-0 kubenswrapper[4752]: E1205 10:36:27.956028 4752 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:27.956129 master-0 kubenswrapper[4752]: E1205 10:36:27.956090 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:36:35.956072609 +0000 UTC m=+77.496195938 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:29.795871 master-0 kubenswrapper[4752]: I1205 10:36:29.795799 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:29.796387 master-0 kubenswrapper[4752]: E1205 10:36:29.795957 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:30.805817 master-0 kubenswrapper[4752]: W1205 10:36:30.805714 4752 warnings.go:70] would violate PodSecurity "restricted:latest": host namespaces (hostNetwork=true), hostPort (container "etcd" uses hostPorts 2379, 2380), privileged (containers "etcdctl", "etcd" must not set securityContext.privileged=true), allowPrivilegeEscalation != false (containers "etcdctl", "etcd" must set securityContext.allowPrivilegeEscalation=false), unrestricted capabilities (containers "etcdctl", "etcd" must set securityContext.capabilities.drop=["ALL"]), restricted volume types (volumes "certs", "data-dir" use restricted volume type "hostPath"), runAsNonRoot != true (pod or containers "etcdctl", "etcd" must set securityContext.runAsNonRoot=true), seccompProfile (pod or containers "etcdctl", "etcd" must set securityContext.seccompProfile.type to "RuntimeDefault" or "Localhost") Dec 05 10:36:30.806809 master-0 kubenswrapper[4752]: I1205 10:36:30.806699 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-master-0-master-0"] Dec 05 10:36:31.032260 master-0 kubenswrapper[4752]: I1205 10:36:31.032185 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-lxmgz" event={"ID":"6a9f011f-36f1-4308-a365-69425c186c7f","Type":"ContainerStarted","Data":"86e525656761107defcc74923af9ddc22ad133898b5a5bd23a42c24a6f90b258"} Dec 05 10:36:31.037579 master-0 kubenswrapper[4752]: I1205 10:36:31.037454 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dms5d" event={"ID":"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8","Type":"ContainerStarted","Data":"215c293fe494f6f41ad17e025bbeffa977136f630518896e92b337f0d84c340a"} Dec 05 10:36:31.048669 master-0 kubenswrapper[4752]: I1205 10:36:31.048606 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-master-0-master-0" podStartSLOduration=1.048559671 podStartE2EDuration="1.048559671s" podCreationTimestamp="2025-12-05 10:36:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:36:31.048151171 +0000 UTC m=+72.588274530" watchObservedRunningTime="2025-12-05 10:36:31.048559671 +0000 UTC m=+72.588683010" Dec 05 10:36:31.090283 master-0 kubenswrapper[4752]: I1205 10:36:31.090168 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-lxmgz" podStartSLOduration=1.07353386 podStartE2EDuration="12.090145473s" podCreationTimestamp="2025-12-05 10:36:19 +0000 UTC" firstStartedPulling="2025-12-05 10:36:19.816075445 +0000 UTC m=+61.356198814" lastFinishedPulling="2025-12-05 10:36:30.832687098 +0000 UTC m=+72.372810427" observedRunningTime="2025-12-05 10:36:31.071160288 +0000 UTC m=+72.611283637" watchObservedRunningTime="2025-12-05 10:36:31.090145473 +0000 UTC m=+72.630268822" Dec 05 10:36:31.581250 master-0 kubenswrapper[4752]: I1205 10:36:31.581126 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:36:31.581250 master-0 kubenswrapper[4752]: E1205 10:36:31.581277 4752 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:31.581822 master-0 kubenswrapper[4752]: E1205 10:36:31.581348 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:03.581325278 +0000 UTC m=+105.121448607 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:36:31.617989 master-0 kubenswrapper[4752]: I1205 10:36:31.617917 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp"] Dec 05 10:36:31.618357 master-0 kubenswrapper[4752]: I1205 10:36:31.618236 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.620603 master-0 kubenswrapper[4752]: I1205 10:36:31.620509 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 10:36:31.620953 master-0 kubenswrapper[4752]: I1205 10:36:31.620803 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 10:36:31.620953 master-0 kubenswrapper[4752]: I1205 10:36:31.620920 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 10:36:31.621794 master-0 kubenswrapper[4752]: I1205 10:36:31.621732 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 10:36:31.622239 master-0 kubenswrapper[4752]: I1205 10:36:31.622191 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 10:36:31.682328 master-0 kubenswrapper[4752]: I1205 10:36:31.682249 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b760849c-8d83-47da-8677-68445c143bef-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.682556 master-0 kubenswrapper[4752]: I1205 10:36:31.682343 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-ovnkube-config\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.682556 master-0 kubenswrapper[4752]: I1205 10:36:31.682380 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-env-overrides\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.682654 master-0 kubenswrapper[4752]: I1205 10:36:31.682578 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfz87\" (UniqueName: \"kubernetes.io/projected/b760849c-8d83-47da-8677-68445c143bef-kube-api-access-jfz87\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.782936 master-0 kubenswrapper[4752]: I1205 10:36:31.782881 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-ovnkube-config\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.783142 master-0 kubenswrapper[4752]: I1205 10:36:31.783085 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-env-overrides\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.783142 master-0 kubenswrapper[4752]: I1205 10:36:31.783132 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfz87\" (UniqueName: \"kubernetes.io/projected/b760849c-8d83-47da-8677-68445c143bef-kube-api-access-jfz87\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.784122 master-0 kubenswrapper[4752]: I1205 10:36:31.783368 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b760849c-8d83-47da-8677-68445c143bef-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.784122 master-0 kubenswrapper[4752]: I1205 10:36:31.783769 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-env-overrides\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.784332 master-0 kubenswrapper[4752]: I1205 10:36:31.784276 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-ovnkube-config\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.788221 master-0 kubenswrapper[4752]: I1205 10:36:31.788182 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b760849c-8d83-47da-8677-68445c143bef-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.788560 master-0 kubenswrapper[4752]: I1205 10:36:31.788535 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-d89ht"] Dec 05 10:36:31.789319 master-0 kubenswrapper[4752]: I1205 10:36:31.789289 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.791187 master-0 kubenswrapper[4752]: I1205 10:36:31.790999 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 10:36:31.791529 master-0 kubenswrapper[4752]: I1205 10:36:31.791486 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 10:36:31.795597 master-0 kubenswrapper[4752]: I1205 10:36:31.795551 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:31.795757 master-0 kubenswrapper[4752]: E1205 10:36:31.795653 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:31.810758 master-0 kubenswrapper[4752]: I1205 10:36:31.810697 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfz87\" (UniqueName: \"kubernetes.io/projected/b760849c-8d83-47da-8677-68445c143bef-kube-api-access-jfz87\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.883938 master-0 kubenswrapper[4752]: I1205 10:36:31.883811 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-bin\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.883938 master-0 kubenswrapper[4752]: I1205 10:36:31.883878 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnxbz\" (UniqueName: \"kubernetes.io/projected/b2a62edc-8f1c-43be-b355-8484bbebac86-kube-api-access-vnxbz\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.883938 master-0 kubenswrapper[4752]: I1205 10:36:31.883905 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-kubelet\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.883938 master-0 kubenswrapper[4752]: I1205 10:36:31.883929 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-node-log\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884166 master-0 kubenswrapper[4752]: I1205 10:36:31.883954 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-script-lib\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884166 master-0 kubenswrapper[4752]: I1205 10:36:31.883975 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-netd\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884166 master-0 kubenswrapper[4752]: I1205 10:36:31.884111 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-openvswitch\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884285 master-0 kubenswrapper[4752]: I1205 10:36:31.884173 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884285 master-0 kubenswrapper[4752]: I1205 10:36:31.884201 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-config\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884285 master-0 kubenswrapper[4752]: I1205 10:36:31.884252 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-ovn-kubernetes\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884398 master-0 kubenswrapper[4752]: I1205 10:36:31.884326 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-systemd\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884472 master-0 kubenswrapper[4752]: I1205 10:36:31.884400 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-netns\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884517 master-0 kubenswrapper[4752]: I1205 10:36:31.884466 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-log-socket\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884517 master-0 kubenswrapper[4752]: I1205 10:36:31.884504 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b2a62edc-8f1c-43be-b355-8484bbebac86-ovn-node-metrics-cert\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884594 master-0 kubenswrapper[4752]: I1205 10:36:31.884541 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-var-lib-openvswitch\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884594 master-0 kubenswrapper[4752]: I1205 10:36:31.884581 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-etc-openvswitch\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884664 master-0 kubenswrapper[4752]: I1205 10:36:31.884613 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-systemd-units\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884664 master-0 kubenswrapper[4752]: I1205 10:36:31.884643 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-slash\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884738 master-0 kubenswrapper[4752]: I1205 10:36:31.884677 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-ovn\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.884738 master-0 kubenswrapper[4752]: I1205 10:36:31.884722 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-env-overrides\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.932128 master-0 kubenswrapper[4752]: I1205 10:36:31.932067 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:36:31.944153 master-0 kubenswrapper[4752]: W1205 10:36:31.944074 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb760849c_8d83_47da_8677_68445c143bef.slice/crio-6607da6ec4214a3fe2d315e70d9756b15e987231bb8310517d849da88c8e7b15 WatchSource:0}: Error finding container 6607da6ec4214a3fe2d315e70d9756b15e987231bb8310517d849da88c8e7b15: Status 404 returned error can't find the container with id 6607da6ec4214a3fe2d315e70d9756b15e987231bb8310517d849da88c8e7b15 Dec 05 10:36:31.986006 master-0 kubenswrapper[4752]: I1205 10:36:31.985944 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-kubelet\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986197 master-0 kubenswrapper[4752]: I1205 10:36:31.986067 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-kubelet\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986197 master-0 kubenswrapper[4752]: I1205 10:36:31.986088 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-node-log\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986197 master-0 kubenswrapper[4752]: I1205 10:36:31.986066 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-node-log\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986197 master-0 kubenswrapper[4752]: I1205 10:36:31.986142 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-script-lib\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986197 master-0 kubenswrapper[4752]: I1205 10:36:31.986174 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-netd\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986348 master-0 kubenswrapper[4752]: I1205 10:36:31.986208 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986348 master-0 kubenswrapper[4752]: I1205 10:36:31.986239 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-config\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986348 master-0 kubenswrapper[4752]: I1205 10:36:31.986290 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986348 master-0 kubenswrapper[4752]: I1205 10:36:31.986284 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-netd\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986549 master-0 kubenswrapper[4752]: I1205 10:36:31.986395 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-openvswitch\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986549 master-0 kubenswrapper[4752]: I1205 10:36:31.986468 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-openvswitch\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986549 master-0 kubenswrapper[4752]: I1205 10:36:31.986516 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-systemd\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986628 master-0 kubenswrapper[4752]: I1205 10:36:31.986570 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-ovn-kubernetes\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986628 master-0 kubenswrapper[4752]: I1205 10:36:31.986611 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-systemd\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986696 master-0 kubenswrapper[4752]: I1205 10:36:31.986623 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-netns\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986696 master-0 kubenswrapper[4752]: I1205 10:36:31.986662 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-netns\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986696 master-0 kubenswrapper[4752]: I1205 10:36:31.986669 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-log-socket\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986801 master-0 kubenswrapper[4752]: I1205 10:36:31.986713 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-log-socket\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986801 master-0 kubenswrapper[4752]: I1205 10:36:31.986626 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-ovn-kubernetes\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986801 master-0 kubenswrapper[4752]: I1205 10:36:31.986724 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b2a62edc-8f1c-43be-b355-8484bbebac86-ovn-node-metrics-cert\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986801 master-0 kubenswrapper[4752]: I1205 10:36:31.986769 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-var-lib-openvswitch\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986906 master-0 kubenswrapper[4752]: I1205 10:36:31.986799 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-systemd-units\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986906 master-0 kubenswrapper[4752]: I1205 10:36:31.986842 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-var-lib-openvswitch\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986906 master-0 kubenswrapper[4752]: I1205 10:36:31.986857 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-systemd-units\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.986906 master-0 kubenswrapper[4752]: I1205 10:36:31.986884 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-etc-openvswitch\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987013 master-0 kubenswrapper[4752]: I1205 10:36:31.986928 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-slash\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987013 master-0 kubenswrapper[4752]: I1205 10:36:31.986985 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-slash\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987013 master-0 kubenswrapper[4752]: I1205 10:36:31.986984 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-etc-openvswitch\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987096 master-0 kubenswrapper[4752]: I1205 10:36:31.987014 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-ovn\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987096 master-0 kubenswrapper[4752]: I1205 10:36:31.987051 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-ovn\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987096 master-0 kubenswrapper[4752]: I1205 10:36:31.987058 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-env-overrides\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987174 master-0 kubenswrapper[4752]: I1205 10:36:31.987093 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-bin\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987174 master-0 kubenswrapper[4752]: I1205 10:36:31.987124 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnxbz\" (UniqueName: \"kubernetes.io/projected/b2a62edc-8f1c-43be-b355-8484bbebac86-kube-api-access-vnxbz\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987272 master-0 kubenswrapper[4752]: I1205 10:36:31.987232 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-bin\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987612 master-0 kubenswrapper[4752]: I1205 10:36:31.987585 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-config\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987855 master-0 kubenswrapper[4752]: I1205 10:36:31.987828 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-env-overrides\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.987905 master-0 kubenswrapper[4752]: I1205 10:36:31.987854 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-script-lib\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:31.990595 master-0 kubenswrapper[4752]: I1205 10:36:31.990563 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b2a62edc-8f1c-43be-b355-8484bbebac86-ovn-node-metrics-cert\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:32.007923 master-0 kubenswrapper[4752]: I1205 10:36:32.007877 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnxbz\" (UniqueName: \"kubernetes.io/projected/b2a62edc-8f1c-43be-b355-8484bbebac86-kube-api-access-vnxbz\") pod \"ovnkube-node-d89ht\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:32.041343 master-0 kubenswrapper[4752]: I1205 10:36:32.041299 4752 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="215c293fe494f6f41ad17e025bbeffa977136f630518896e92b337f0d84c340a" exitCode=0 Dec 05 10:36:32.041557 master-0 kubenswrapper[4752]: I1205 10:36:32.041375 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dms5d" event={"ID":"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8","Type":"ContainerDied","Data":"215c293fe494f6f41ad17e025bbeffa977136f630518896e92b337f0d84c340a"} Dec 05 10:36:32.043666 master-0 kubenswrapper[4752]: I1205 10:36:32.043215 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" event={"ID":"b760849c-8d83-47da-8677-68445c143bef","Type":"ContainerStarted","Data":"6607da6ec4214a3fe2d315e70d9756b15e987231bb8310517d849da88c8e7b15"} Dec 05 10:36:32.131994 master-0 kubenswrapper[4752]: I1205 10:36:32.131896 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:36:32.141190 master-0 kubenswrapper[4752]: W1205 10:36:32.141142 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2a62edc_8f1c_43be_b355_8484bbebac86.slice/crio-a20c01724bf2a194e055334e233e74f0b08a60cb52f191783b538ff245ddd1cb WatchSource:0}: Error finding container a20c01724bf2a194e055334e233e74f0b08a60cb52f191783b538ff245ddd1cb: Status 404 returned error can't find the container with id a20c01724bf2a194e055334e233e74f0b08a60cb52f191783b538ff245ddd1cb Dec 05 10:36:33.048637 master-0 kubenswrapper[4752]: I1205 10:36:33.048571 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" event={"ID":"b760849c-8d83-47da-8677-68445c143bef","Type":"ContainerStarted","Data":"b6da3037c684f1d555a5e86bd43c09ae597015960dfb7128064a2e9e65f458b4"} Dec 05 10:36:33.050303 master-0 kubenswrapper[4752]: I1205 10:36:33.050276 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerStarted","Data":"a20c01724bf2a194e055334e233e74f0b08a60cb52f191783b538ff245ddd1cb"} Dec 05 10:36:33.053230 master-0 kubenswrapper[4752]: I1205 10:36:33.053176 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dms5d" event={"ID":"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8","Type":"ContainerStarted","Data":"f4f209477f798940668277c240fd1326e8fd7d7c2b8eedd2830e59be706883e8"} Dec 05 10:36:33.795784 master-0 kubenswrapper[4752]: I1205 10:36:33.795335 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:33.795994 master-0 kubenswrapper[4752]: E1205 10:36:33.795916 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:34.057877 master-0 kubenswrapper[4752]: I1205 10:36:34.057752 4752 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="f4f209477f798940668277c240fd1326e8fd7d7c2b8eedd2830e59be706883e8" exitCode=0 Dec 05 10:36:34.057877 master-0 kubenswrapper[4752]: I1205 10:36:34.057791 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dms5d" event={"ID":"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8","Type":"ContainerDied","Data":"f4f209477f798940668277c240fd1326e8fd7d7c2b8eedd2830e59be706883e8"} Dec 05 10:36:34.771971 master-0 kubenswrapper[4752]: I1205 10:36:34.771925 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-d6fzk"] Dec 05 10:36:34.772285 master-0 kubenswrapper[4752]: I1205 10:36:34.772260 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:34.772416 master-0 kubenswrapper[4752]: E1205 10:36:34.772329 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:34.810404 master-0 kubenswrapper[4752]: I1205 10:36:34.810333 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:34.911490 master-0 kubenswrapper[4752]: I1205 10:36:34.911405 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:34.924974 master-0 kubenswrapper[4752]: E1205 10:36:34.924912 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 10:36:34.924974 master-0 kubenswrapper[4752]: E1205 10:36:34.924961 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 10:36:34.924974 master-0 kubenswrapper[4752]: E1205 10:36:34.924978 4752 projected.go:194] Error preparing data for projected volume kube-api-access-5n7tf for pod openshift-network-diagnostics/network-check-target-d6fzk: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:34.925288 master-0 kubenswrapper[4752]: E1205 10:36:34.925109 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf podName:720a7467-ce93-4d48-82ec-9ad0922d99c2 nodeName:}" failed. No retries permitted until 2025-12-05 10:36:35.425049148 +0000 UTC m=+76.965172517 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-5n7tf" (UniqueName: "kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf") pod "network-check-target-d6fzk" (UID: "720a7467-ce93-4d48-82ec-9ad0922d99c2") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:35.516628 master-0 kubenswrapper[4752]: I1205 10:36:35.516534 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:35.517173 master-0 kubenswrapper[4752]: E1205 10:36:35.516769 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 10:36:35.517173 master-0 kubenswrapper[4752]: E1205 10:36:35.516813 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 10:36:35.517173 master-0 kubenswrapper[4752]: E1205 10:36:35.516831 4752 projected.go:194] Error preparing data for projected volume kube-api-access-5n7tf for pod openshift-network-diagnostics/network-check-target-d6fzk: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:35.517173 master-0 kubenswrapper[4752]: E1205 10:36:35.516907 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf podName:720a7467-ce93-4d48-82ec-9ad0922d99c2 nodeName:}" failed. No retries permitted until 2025-12-05 10:36:36.516881583 +0000 UTC m=+78.057004982 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-5n7tf" (UniqueName: "kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf") pod "network-check-target-d6fzk" (UID: "720a7467-ce93-4d48-82ec-9ad0922d99c2") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:35.794996 master-0 kubenswrapper[4752]: I1205 10:36:35.794941 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:35.795223 master-0 kubenswrapper[4752]: E1205 10:36:35.795063 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:36.020121 master-0 kubenswrapper[4752]: I1205 10:36:36.020058 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:36.020894 master-0 kubenswrapper[4752]: E1205 10:36:36.020832 4752 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:36.021016 master-0 kubenswrapper[4752]: E1205 10:36:36.020991 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:36:52.020952027 +0000 UTC m=+93.561075356 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:36.069354 master-0 kubenswrapper[4752]: I1205 10:36:36.069237 4752 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="3acdc4bc20dace9999e3e38bd13059d0977882b304881b0e3ccd402e31f993c2" exitCode=0 Dec 05 10:36:36.069354 master-0 kubenswrapper[4752]: I1205 10:36:36.069281 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dms5d" event={"ID":"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8","Type":"ContainerDied","Data":"3acdc4bc20dace9999e3e38bd13059d0977882b304881b0e3ccd402e31f993c2"} Dec 05 10:36:36.524988 master-0 kubenswrapper[4752]: I1205 10:36:36.524870 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:36.525476 master-0 kubenswrapper[4752]: E1205 10:36:36.525035 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 10:36:36.525476 master-0 kubenswrapper[4752]: E1205 10:36:36.525062 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 10:36:36.525476 master-0 kubenswrapper[4752]: E1205 10:36:36.525073 4752 projected.go:194] Error preparing data for projected volume kube-api-access-5n7tf for pod openshift-network-diagnostics/network-check-target-d6fzk: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:36.525476 master-0 kubenswrapper[4752]: E1205 10:36:36.525117 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf podName:720a7467-ce93-4d48-82ec-9ad0922d99c2 nodeName:}" failed. No retries permitted until 2025-12-05 10:36:38.525100514 +0000 UTC m=+80.065223843 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-5n7tf" (UniqueName: "kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf") pod "network-check-target-d6fzk" (UID: "720a7467-ce93-4d48-82ec-9ad0922d99c2") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:36.795598 master-0 kubenswrapper[4752]: I1205 10:36:36.795559 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:36.795774 master-0 kubenswrapper[4752]: E1205 10:36:36.795690 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:37.384477 master-0 kubenswrapper[4752]: I1205 10:36:37.384234 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-ql7j7"] Dec 05 10:36:37.384843 master-0 kubenswrapper[4752]: I1205 10:36:37.384804 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.386684 master-0 kubenswrapper[4752]: I1205 10:36:37.386181 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 10:36:37.388293 master-0 kubenswrapper[4752]: I1205 10:36:37.387560 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 10:36:37.388293 master-0 kubenswrapper[4752]: I1205 10:36:37.387758 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 10:36:37.388293 master-0 kubenswrapper[4752]: I1205 10:36:37.387903 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 10:36:37.388787 master-0 kubenswrapper[4752]: I1205 10:36:37.388638 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 10:36:37.431139 master-0 kubenswrapper[4752]: I1205 10:36:37.431041 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-ovnkube-identity-cm\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.431338 master-0 kubenswrapper[4752]: I1205 10:36:37.431259 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-webhook-cert\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.431338 master-0 kubenswrapper[4752]: I1205 10:36:37.431309 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg9zq\" (UniqueName: \"kubernetes.io/projected/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-kube-api-access-wg9zq\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.431405 master-0 kubenswrapper[4752]: I1205 10:36:37.431352 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-env-overrides\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.532546 master-0 kubenswrapper[4752]: I1205 10:36:37.532478 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-env-overrides\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.533326 master-0 kubenswrapper[4752]: I1205 10:36:37.532849 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-ovnkube-identity-cm\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.533326 master-0 kubenswrapper[4752]: I1205 10:36:37.532965 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-webhook-cert\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.533326 master-0 kubenswrapper[4752]: I1205 10:36:37.532991 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg9zq\" (UniqueName: \"kubernetes.io/projected/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-kube-api-access-wg9zq\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.533573 master-0 kubenswrapper[4752]: I1205 10:36:37.533535 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-env-overrides\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.535931 master-0 kubenswrapper[4752]: I1205 10:36:37.535892 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-ovnkube-identity-cm\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.541487 master-0 kubenswrapper[4752]: I1205 10:36:37.537628 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-webhook-cert\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.552396 master-0 kubenswrapper[4752]: I1205 10:36:37.552254 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg9zq\" (UniqueName: \"kubernetes.io/projected/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-kube-api-access-wg9zq\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.707970 master-0 kubenswrapper[4752]: I1205 10:36:37.707841 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:36:37.795760 master-0 kubenswrapper[4752]: I1205 10:36:37.795709 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:37.795947 master-0 kubenswrapper[4752]: E1205 10:36:37.795870 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:38.076667 master-0 kubenswrapper[4752]: I1205 10:36:38.076613 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-ql7j7" event={"ID":"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e","Type":"ContainerStarted","Data":"6ead10efe7a643d9b0dd883cc19f8ef852a5658bc79c235b7c7c6b5de2e97811"} Dec 05 10:36:38.543099 master-0 kubenswrapper[4752]: I1205 10:36:38.543061 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:38.543668 master-0 kubenswrapper[4752]: E1205 10:36:38.543242 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 10:36:38.543668 master-0 kubenswrapper[4752]: E1205 10:36:38.543261 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 10:36:38.543668 master-0 kubenswrapper[4752]: E1205 10:36:38.543273 4752 projected.go:194] Error preparing data for projected volume kube-api-access-5n7tf for pod openshift-network-diagnostics/network-check-target-d6fzk: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:38.543668 master-0 kubenswrapper[4752]: E1205 10:36:38.543359 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf podName:720a7467-ce93-4d48-82ec-9ad0922d99c2 nodeName:}" failed. No retries permitted until 2025-12-05 10:36:42.543309979 +0000 UTC m=+84.083433308 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-5n7tf" (UniqueName: "kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf") pod "network-check-target-d6fzk" (UID: "720a7467-ce93-4d48-82ec-9ad0922d99c2") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:38.796729 master-0 kubenswrapper[4752]: I1205 10:36:38.796623 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:38.796729 master-0 kubenswrapper[4752]: E1205 10:36:38.796714 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:39.795411 master-0 kubenswrapper[4752]: I1205 10:36:39.795346 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:39.796081 master-0 kubenswrapper[4752]: E1205 10:36:39.795499 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:40.795589 master-0 kubenswrapper[4752]: I1205 10:36:40.795513 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:40.796121 master-0 kubenswrapper[4752]: E1205 10:36:40.795663 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:41.795713 master-0 kubenswrapper[4752]: I1205 10:36:41.795612 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:41.796186 master-0 kubenswrapper[4752]: E1205 10:36:41.795905 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:42.580132 master-0 kubenswrapper[4752]: I1205 10:36:42.580047 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:42.580351 master-0 kubenswrapper[4752]: E1205 10:36:42.580286 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 10:36:42.580351 master-0 kubenswrapper[4752]: E1205 10:36:42.580323 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 10:36:42.580351 master-0 kubenswrapper[4752]: E1205 10:36:42.580336 4752 projected.go:194] Error preparing data for projected volume kube-api-access-5n7tf for pod openshift-network-diagnostics/network-check-target-d6fzk: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:42.580626 master-0 kubenswrapper[4752]: E1205 10:36:42.580404 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf podName:720a7467-ce93-4d48-82ec-9ad0922d99c2 nodeName:}" failed. No retries permitted until 2025-12-05 10:36:50.580384226 +0000 UTC m=+92.120507575 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-5n7tf" (UniqueName: "kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf") pod "network-check-target-d6fzk" (UID: "720a7467-ce93-4d48-82ec-9ad0922d99c2") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:42.796017 master-0 kubenswrapper[4752]: I1205 10:36:42.795950 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:42.796478 master-0 kubenswrapper[4752]: E1205 10:36:42.796091 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:43.795761 master-0 kubenswrapper[4752]: I1205 10:36:43.795598 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:43.795983 master-0 kubenswrapper[4752]: E1205 10:36:43.795931 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:44.795253 master-0 kubenswrapper[4752]: I1205 10:36:44.795185 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:44.795884 master-0 kubenswrapper[4752]: E1205 10:36:44.795309 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:45.795628 master-0 kubenswrapper[4752]: I1205 10:36:45.795555 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:45.796181 master-0 kubenswrapper[4752]: E1205 10:36:45.795716 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:46.795749 master-0 kubenswrapper[4752]: I1205 10:36:46.795674 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:46.796601 master-0 kubenswrapper[4752]: E1205 10:36:46.795825 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:47.795256 master-0 kubenswrapper[4752]: I1205 10:36:47.795145 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:47.795556 master-0 kubenswrapper[4752]: E1205 10:36:47.795309 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:48.795412 master-0 kubenswrapper[4752]: I1205 10:36:48.795308 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:48.797043 master-0 kubenswrapper[4752]: E1205 10:36:48.796936 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:49.795621 master-0 kubenswrapper[4752]: I1205 10:36:49.795561 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:49.796222 master-0 kubenswrapper[4752]: E1205 10:36:49.795691 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:50.027325 master-0 kubenswrapper[4752]: I1205 10:36:50.027269 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/bootstrap-kube-apiserver-master-0"] Dec 05 10:36:50.652061 master-0 kubenswrapper[4752]: I1205 10:36:50.651984 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:50.652268 master-0 kubenswrapper[4752]: E1205 10:36:50.652206 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 10:36:50.652268 master-0 kubenswrapper[4752]: E1205 10:36:50.652243 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 10:36:50.652268 master-0 kubenswrapper[4752]: E1205 10:36:50.652259 4752 projected.go:194] Error preparing data for projected volume kube-api-access-5n7tf for pod openshift-network-diagnostics/network-check-target-d6fzk: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:50.652357 master-0 kubenswrapper[4752]: E1205 10:36:50.652325 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf podName:720a7467-ce93-4d48-82ec-9ad0922d99c2 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:06.652304886 +0000 UTC m=+108.192428225 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-5n7tf" (UniqueName: "kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf") pod "network-check-target-d6fzk" (UID: "720a7467-ce93-4d48-82ec-9ad0922d99c2") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:36:50.795926 master-0 kubenswrapper[4752]: I1205 10:36:50.795811 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:50.795926 master-0 kubenswrapper[4752]: E1205 10:36:50.795936 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:51.795550 master-0 kubenswrapper[4752]: I1205 10:36:51.795481 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:51.795762 master-0 kubenswrapper[4752]: E1205 10:36:51.795620 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:52.065269 master-0 kubenswrapper[4752]: I1205 10:36:52.065099 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:52.066236 master-0 kubenswrapper[4752]: E1205 10:36:52.065360 4752 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:52.066236 master-0 kubenswrapper[4752]: E1205 10:36:52.065477 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:37:24.065452658 +0000 UTC m=+125.605576007 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 10:36:52.795879 master-0 kubenswrapper[4752]: I1205 10:36:52.795829 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:52.796125 master-0 kubenswrapper[4752]: E1205 10:36:52.796046 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:52.808885 master-0 kubenswrapper[4752]: I1205 10:36:52.808825 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["kube-system/bootstrap-kube-controller-manager-master-0"] Dec 05 10:36:53.795140 master-0 kubenswrapper[4752]: I1205 10:36:53.795098 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:53.795738 master-0 kubenswrapper[4752]: E1205 10:36:53.795212 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:54.795314 master-0 kubenswrapper[4752]: I1205 10:36:54.795245 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:54.796199 master-0 kubenswrapper[4752]: E1205 10:36:54.795383 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:55.124818 master-0 kubenswrapper[4752]: I1205 10:36:55.124726 4752 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="ee9e5b75d34cda6a745ee4f00c4f147db05bd9943033e03d413d4fa85d63bb34" exitCode=0 Dec 05 10:36:55.125055 master-0 kubenswrapper[4752]: I1205 10:36:55.124824 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dms5d" event={"ID":"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8","Type":"ContainerDied","Data":"ee9e5b75d34cda6a745ee4f00c4f147db05bd9943033e03d413d4fa85d63bb34"} Dec 05 10:36:55.127449 master-0 kubenswrapper[4752]: I1205 10:36:55.127267 4752 generic.go:334] "Generic (PLEG): container finished" podID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerID="3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f" exitCode=0 Dec 05 10:36:55.127449 master-0 kubenswrapper[4752]: I1205 10:36:55.127386 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerDied","Data":"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f"} Dec 05 10:36:55.131505 master-0 kubenswrapper[4752]: I1205 10:36:55.131414 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-ql7j7" event={"ID":"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e","Type":"ContainerStarted","Data":"5674d03e599731519f29f243f076ae28159025706182191e04e4bbfef0819512"} Dec 05 10:36:55.131505 master-0 kubenswrapper[4752]: I1205 10:36:55.131508 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-ql7j7" event={"ID":"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e","Type":"ContainerStarted","Data":"a7e1b69ed43ea60bd1d549f53fbdffb3c63ba4aa5f33456e1b6e9ca137681118"} Dec 05 10:36:55.134983 master-0 kubenswrapper[4752]: I1205 10:36:55.134377 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" event={"ID":"b760849c-8d83-47da-8677-68445c143bef","Type":"ContainerStarted","Data":"d5454b7ce1bc247671ab46448edd5a6f0a198a1673e48f822fea3525f6db868a"} Dec 05 10:36:55.190248 master-0 kubenswrapper[4752]: I1205 10:36:55.190168 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" podStartSLOduration=5.190148013 podStartE2EDuration="5.190148013s" podCreationTimestamp="2025-12-05 10:36:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:36:55.189824355 +0000 UTC m=+96.729947714" watchObservedRunningTime="2025-12-05 10:36:55.190148013 +0000 UTC m=+96.730271342" Dec 05 10:36:55.210706 master-0 kubenswrapper[4752]: I1205 10:36:55.210495 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/bootstrap-kube-controller-manager-master-0" podStartSLOduration=3.210476012 podStartE2EDuration="3.210476012s" podCreationTimestamp="2025-12-05 10:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:36:55.205544576 +0000 UTC m=+96.745667905" watchObservedRunningTime="2025-12-05 10:36:55.210476012 +0000 UTC m=+96.750599361" Dec 05 10:36:55.271871 master-0 kubenswrapper[4752]: I1205 10:36:55.271770 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" podStartSLOduration=2.179958087 podStartE2EDuration="24.271750767s" podCreationTimestamp="2025-12-05 10:36:31 +0000 UTC" firstStartedPulling="2025-12-05 10:36:32.122715295 +0000 UTC m=+73.662838634" lastFinishedPulling="2025-12-05 10:36:54.214507955 +0000 UTC m=+95.754631314" observedRunningTime="2025-12-05 10:36:55.271530031 +0000 UTC m=+96.811653380" watchObservedRunningTime="2025-12-05 10:36:55.271750767 +0000 UTC m=+96.811874106" Dec 05 10:36:55.358667 master-0 kubenswrapper[4752]: I1205 10:36:55.358553 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-node-identity/network-node-identity-ql7j7" podStartSLOduration=1.826947783 podStartE2EDuration="18.358534624s" podCreationTimestamp="2025-12-05 10:36:37 +0000 UTC" firstStartedPulling="2025-12-05 10:36:37.720951816 +0000 UTC m=+79.261075145" lastFinishedPulling="2025-12-05 10:36:54.252538657 +0000 UTC m=+95.792661986" observedRunningTime="2025-12-05 10:36:55.358098132 +0000 UTC m=+96.898221471" watchObservedRunningTime="2025-12-05 10:36:55.358534624 +0000 UTC m=+96.898657953" Dec 05 10:36:55.795067 master-0 kubenswrapper[4752]: I1205 10:36:55.795008 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:55.795263 master-0 kubenswrapper[4752]: E1205 10:36:55.795155 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:56.144753 master-0 kubenswrapper[4752]: I1205 10:36:56.144340 4752 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="12675055319a88464b4e8137b9c8f20eb9ba5b578bde42cad92050f544a0c6be" exitCode=0 Dec 05 10:36:56.146106 master-0 kubenswrapper[4752]: I1205 10:36:56.144494 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dms5d" event={"ID":"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8","Type":"ContainerDied","Data":"12675055319a88464b4e8137b9c8f20eb9ba5b578bde42cad92050f544a0c6be"} Dec 05 10:36:56.150973 master-0 kubenswrapper[4752]: I1205 10:36:56.150868 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerStarted","Data":"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87"} Dec 05 10:36:56.150973 master-0 kubenswrapper[4752]: I1205 10:36:56.150970 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerStarted","Data":"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3"} Dec 05 10:36:56.151218 master-0 kubenswrapper[4752]: I1205 10:36:56.150992 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerStarted","Data":"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049"} Dec 05 10:36:56.151218 master-0 kubenswrapper[4752]: I1205 10:36:56.151013 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerStarted","Data":"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5"} Dec 05 10:36:56.151218 master-0 kubenswrapper[4752]: I1205 10:36:56.151034 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerStarted","Data":"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2"} Dec 05 10:36:56.151218 master-0 kubenswrapper[4752]: I1205 10:36:56.151053 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerStarted","Data":"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e"} Dec 05 10:36:56.796311 master-0 kubenswrapper[4752]: I1205 10:36:56.795736 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:56.796311 master-0 kubenswrapper[4752]: E1205 10:36:56.795891 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:56.862304 master-0 kubenswrapper[4752]: I1205 10:36:56.862237 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["kube-system/bootstrap-kube-scheduler-master-0"] Dec 05 10:36:57.160346 master-0 kubenswrapper[4752]: I1205 10:36:57.160121 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-dms5d" event={"ID":"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8","Type":"ContainerStarted","Data":"138bb4f48ce80044b27dcfc412d9faaf143e50a6bdaf00f64f8fc52602972c16"} Dec 05 10:36:57.197285 master-0 kubenswrapper[4752]: I1205 10:36:57.197190 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-dms5d" podStartSLOduration=3.800805155 podStartE2EDuration="38.197166212s" podCreationTimestamp="2025-12-05 10:36:19 +0000 UTC" firstStartedPulling="2025-12-05 10:36:19.732222583 +0000 UTC m=+61.272345902" lastFinishedPulling="2025-12-05 10:36:54.12858363 +0000 UTC m=+95.668706959" observedRunningTime="2025-12-05 10:36:57.197003838 +0000 UTC m=+98.737127227" watchObservedRunningTime="2025-12-05 10:36:57.197166212 +0000 UTC m=+98.737289571" Dec 05 10:36:57.197932 master-0 kubenswrapper[4752]: I1205 10:36:57.197872 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/bootstrap-kube-scheduler-master-0" podStartSLOduration=1.197865769 podStartE2EDuration="1.197865769s" podCreationTimestamp="2025-12-05 10:36:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:36:57.173727783 +0000 UTC m=+98.713851152" watchObservedRunningTime="2025-12-05 10:36:57.197865769 +0000 UTC m=+98.737989108" Dec 05 10:36:57.795013 master-0 kubenswrapper[4752]: I1205 10:36:57.794931 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:57.795461 master-0 kubenswrapper[4752]: E1205 10:36:57.795062 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:36:58.169511 master-0 kubenswrapper[4752]: I1205 10:36:58.169335 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerStarted","Data":"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98"} Dec 05 10:36:58.795171 master-0 kubenswrapper[4752]: I1205 10:36:58.795080 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:36:58.796928 master-0 kubenswrapper[4752]: E1205 10:36:58.796844 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:36:59.795055 master-0 kubenswrapper[4752]: I1205 10:36:59.794968 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:36:59.795842 master-0 kubenswrapper[4752]: E1205 10:36:59.795152 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:37:00.796059 master-0 kubenswrapper[4752]: I1205 10:37:00.795706 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:00.796761 master-0 kubenswrapper[4752]: E1205 10:37:00.796135 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:37:01.186738 master-0 kubenswrapper[4752]: I1205 10:37:01.186539 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerStarted","Data":"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b"} Dec 05 10:37:01.187071 master-0 kubenswrapper[4752]: I1205 10:37:01.186989 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:37:01.217516 master-0 kubenswrapper[4752]: I1205 10:37:01.217418 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:37:01.219810 master-0 kubenswrapper[4752]: I1205 10:37:01.219736 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" podStartSLOduration=8.128207304 podStartE2EDuration="30.219718587s" podCreationTimestamp="2025-12-05 10:36:31 +0000 UTC" firstStartedPulling="2025-12-05 10:36:32.14250349 +0000 UTC m=+73.682626819" lastFinishedPulling="2025-12-05 10:36:54.234014773 +0000 UTC m=+95.774138102" observedRunningTime="2025-12-05 10:37:01.21941562 +0000 UTC m=+102.759539009" watchObservedRunningTime="2025-12-05 10:37:01.219718587 +0000 UTC m=+102.759841906" Dec 05 10:37:01.795627 master-0 kubenswrapper[4752]: I1205 10:37:01.795569 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:01.795880 master-0 kubenswrapper[4752]: E1205 10:37:01.795795 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:37:02.132633 master-0 kubenswrapper[4752]: I1205 10:37:02.132484 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:37:02.132633 master-0 kubenswrapper[4752]: I1205 10:37:02.132539 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:37:02.153998 master-0 kubenswrapper[4752]: I1205 10:37:02.153950 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:37:02.660928 master-0 kubenswrapper[4752]: I1205 10:37:02.660853 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8gjgm"] Dec 05 10:37:02.661196 master-0 kubenswrapper[4752]: I1205 10:37:02.660961 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:02.661196 master-0 kubenswrapper[4752]: E1205 10:37:02.661055 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:37:02.664392 master-0 kubenswrapper[4752]: I1205 10:37:02.664333 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-d6fzk"] Dec 05 10:37:02.664874 master-0 kubenswrapper[4752]: I1205 10:37:02.664554 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:02.664945 master-0 kubenswrapper[4752]: E1205 10:37:02.664873 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:37:03.665789 master-0 kubenswrapper[4752]: I1205 10:37:03.665720 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:03.666905 master-0 kubenswrapper[4752]: E1205 10:37:03.665904 4752 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:37:03.666905 master-0 kubenswrapper[4752]: E1205 10:37:03.665989 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:38:07.665966674 +0000 UTC m=+169.206090013 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:37:03.795959 master-0 kubenswrapper[4752]: I1205 10:37:03.795807 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:03.795959 master-0 kubenswrapper[4752]: I1205 10:37:03.795912 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:03.796229 master-0 kubenswrapper[4752]: E1205 10:37:03.796063 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:37:03.796229 master-0 kubenswrapper[4752]: E1205 10:37:03.796182 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:37:05.795671 master-0 kubenswrapper[4752]: I1205 10:37:05.795321 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:05.796193 master-0 kubenswrapper[4752]: I1205 10:37:05.795481 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:05.796482 master-0 kubenswrapper[4752]: E1205 10:37:05.796379 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:37:05.796590 master-0 kubenswrapper[4752]: E1205 10:37:05.796540 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:37:06.620513 master-0 kubenswrapper[4752]: I1205 10:37:06.619857 4752 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-d89ht"] Dec 05 10:37:06.620752 master-0 kubenswrapper[4752]: I1205 10:37:06.620692 4752 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovn-controller" containerID="cri-o://21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e" gracePeriod=30 Dec 05 10:37:06.622502 master-0 kubenswrapper[4752]: I1205 10:37:06.620849 4752 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="sbdb" containerID="cri-o://573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98" gracePeriod=30 Dec 05 10:37:06.622502 master-0 kubenswrapper[4752]: I1205 10:37:06.620714 4752 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="nbdb" containerID="cri-o://cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87" gracePeriod=30 Dec 05 10:37:06.622502 master-0 kubenswrapper[4752]: I1205 10:37:06.620940 4752 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="northd" containerID="cri-o://29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3" gracePeriod=30 Dec 05 10:37:06.622502 master-0 kubenswrapper[4752]: I1205 10:37:06.620959 4752 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="kube-rbac-proxy-node" containerID="cri-o://b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5" gracePeriod=30 Dec 05 10:37:06.622502 master-0 kubenswrapper[4752]: I1205 10:37:06.620890 4752 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovn-acl-logging" containerID="cri-o://9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2" gracePeriod=30 Dec 05 10:37:06.622502 master-0 kubenswrapper[4752]: I1205 10:37:06.620946 4752 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049" gracePeriod=30 Dec 05 10:37:06.651599 master-0 kubenswrapper[4752]: I1205 10:37:06.651547 4752 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovnkube-controller" containerID="cri-o://00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b" gracePeriod=30 Dec 05 10:37:06.651823 master-0 kubenswrapper[4752]: I1205 10:37:06.651631 4752 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovnkube-controller" probeResult="failure" output="" Dec 05 10:37:06.693257 master-0 kubenswrapper[4752]: I1205 10:37:06.693208 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:06.693522 master-0 kubenswrapper[4752]: E1205 10:37:06.693353 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 10:37:06.693522 master-0 kubenswrapper[4752]: E1205 10:37:06.693370 4752 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 10:37:06.693522 master-0 kubenswrapper[4752]: E1205 10:37:06.693380 4752 projected.go:194] Error preparing data for projected volume kube-api-access-5n7tf for pod openshift-network-diagnostics/network-check-target-d6fzk: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:37:06.693522 master-0 kubenswrapper[4752]: E1205 10:37:06.693454 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf podName:720a7467-ce93-4d48-82ec-9ad0922d99c2 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:38.693415634 +0000 UTC m=+140.233538963 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-5n7tf" (UniqueName: "kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf") pod "network-check-target-d6fzk" (UID: "720a7467-ce93-4d48-82ec-9ad0922d99c2") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 10:37:06.906863 master-0 kubenswrapper[4752]: I1205 10:37:06.906762 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d89ht_b2a62edc-8f1c-43be-b355-8484bbebac86/ovnkube-controller/0.log" Dec 05 10:37:06.909183 master-0 kubenswrapper[4752]: I1205 10:37:06.909092 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d89ht_b2a62edc-8f1c-43be-b355-8484bbebac86/kube-rbac-proxy-ovn-metrics/0.log" Dec 05 10:37:06.909809 master-0 kubenswrapper[4752]: I1205 10:37:06.909728 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d89ht_b2a62edc-8f1c-43be-b355-8484bbebac86/kube-rbac-proxy-node/0.log" Dec 05 10:37:06.910493 master-0 kubenswrapper[4752]: I1205 10:37:06.910397 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d89ht_b2a62edc-8f1c-43be-b355-8484bbebac86/ovn-acl-logging/0.log" Dec 05 10:37:06.911094 master-0 kubenswrapper[4752]: I1205 10:37:06.911037 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d89ht_b2a62edc-8f1c-43be-b355-8484bbebac86/ovn-controller/0.log" Dec 05 10:37:06.911695 master-0 kubenswrapper[4752]: I1205 10:37:06.911641 4752 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:37:06.960928 master-0 kubenswrapper[4752]: I1205 10:37:06.960828 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rsfjs"] Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: E1205 10:37:06.960953 4752 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="sbdb" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: I1205 10:37:06.960966 4752 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="sbdb" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: E1205 10:37:06.960975 4752 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovnkube-controller" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: I1205 10:37:06.960984 4752 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovnkube-controller" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: E1205 10:37:06.960994 4752 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="northd" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: I1205 10:37:06.961004 4752 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="northd" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: E1205 10:37:06.961013 4752 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="nbdb" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: I1205 10:37:06.961021 4752 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="nbdb" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: E1205 10:37:06.961032 4752 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="kube-rbac-proxy-node" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: I1205 10:37:06.961040 4752 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="kube-rbac-proxy-node" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: E1205 10:37:06.961050 4752 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovn-acl-logging" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: I1205 10:37:06.961060 4752 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovn-acl-logging" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: E1205 10:37:06.961069 4752 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: I1205 10:37:06.961078 4752 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: E1205 10:37:06.961088 4752 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="kubecfg-setup" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: I1205 10:37:06.961096 4752 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="kubecfg-setup" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: E1205 10:37:06.961104 4752 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovn-controller" Dec 05 10:37:06.961126 master-0 kubenswrapper[4752]: I1205 10:37:06.961112 4752 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovn-controller" Dec 05 10:37:06.961793 master-0 kubenswrapper[4752]: I1205 10:37:06.961153 4752 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovn-controller" Dec 05 10:37:06.961793 master-0 kubenswrapper[4752]: I1205 10:37:06.961165 4752 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 10:37:06.961793 master-0 kubenswrapper[4752]: I1205 10:37:06.961177 4752 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovnkube-controller" Dec 05 10:37:06.961793 master-0 kubenswrapper[4752]: I1205 10:37:06.961187 4752 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="ovn-acl-logging" Dec 05 10:37:06.961793 master-0 kubenswrapper[4752]: I1205 10:37:06.961197 4752 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="northd" Dec 05 10:37:06.961793 master-0 kubenswrapper[4752]: I1205 10:37:06.961205 4752 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="nbdb" Dec 05 10:37:06.961793 master-0 kubenswrapper[4752]: I1205 10:37:06.961213 4752 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="kube-rbac-proxy-node" Dec 05 10:37:06.961793 master-0 kubenswrapper[4752]: I1205 10:37:06.961221 4752 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerName="sbdb" Dec 05 10:37:06.962066 master-0 kubenswrapper[4752]: I1205 10:37:06.961989 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.995828 master-0 kubenswrapper[4752]: I1205 10:37:06.995770 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-systemd-units\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.995828 master-0 kubenswrapper[4752]: I1205 10:37:06.995801 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-openvswitch\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.995828 master-0 kubenswrapper[4752]: I1205 10:37:06.995825 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnxbz\" (UniqueName: \"kubernetes.io/projected/b2a62edc-8f1c-43be-b355-8484bbebac86-kube-api-access-vnxbz\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.995828 master-0 kubenswrapper[4752]: I1205 10:37:06.995840 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-slash\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.995860 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-config\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.995878 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-script-lib\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.995893 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-bin\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.995906 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-node-log\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.995893 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.995921 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-ovn-kubernetes\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.995964 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.995964 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-slash" (OuterVolumeSpecName: "host-slash") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.995989 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.996007 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.996040 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-node-log" (OuterVolumeSpecName: "node-log") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.996070 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b2a62edc-8f1c-43be-b355-8484bbebac86-ovn-node-metrics-cert\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.996099 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-systemd\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.996116 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-env-overrides\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.996135 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-netd\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.996156 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-ovn\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996175 master-0 kubenswrapper[4752]: I1205 10:37:06.996172 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-var-lib-cni-networks-ovn-kubernetes\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996192 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-kubelet\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996206 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-netns\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996214 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996221 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-log-socket\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996244 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-log-socket" (OuterVolumeSpecName: "log-socket") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996267 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996277 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-etc-openvswitch\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996288 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996312 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996313 4752 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-var-lib-openvswitch\") pod \"b2a62edc-8f1c-43be-b355-8484bbebac86\" (UID: \"b2a62edc-8f1c-43be-b355-8484bbebac86\") " Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996330 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996393 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996405 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996416 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996482 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-systemd-units\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.996772 master-0 kubenswrapper[4752]: I1205 10:37:06.996507 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-netd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996546 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-netns\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996575 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996597 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-bin\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996622 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-env-overrides\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996622 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996644 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-node-log\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996687 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-systemd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996714 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-config\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996716 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996781 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/495ba1ea-f844-43ec-8be7-47e738f5428a-ovn-node-metrics-cert\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996829 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-script-lib\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996848 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-etc-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996866 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-var-lib-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996867 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996880 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-log-socket\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.997719 master-0 kubenswrapper[4752]: I1205 10:37:06.996990 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997031 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-ovn\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997056 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-kubelet\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997113 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjgc4\" (UniqueName: \"kubernetes.io/projected/495ba1ea-f844-43ec-8be7-47e738f5428a-kube-api-access-tjgc4\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997160 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-slash\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997230 4752 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-var-lib-openvswitch\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997247 4752 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-openvswitch\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997260 4752 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-systemd-units\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997273 4752 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-slash\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997285 4752 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997297 4752 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-ovnkube-script-lib\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997309 4752 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-bin\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997323 4752 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-node-log\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997335 4752 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-ovn-kubernetes\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997349 4752 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-cni-netd\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997361 4752 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b2a62edc-8f1c-43be-b355-8484bbebac86-env-overrides\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997375 4752 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-ovn\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997387 4752 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-var-lib-cni-networks-ovn-kubernetes\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997401 4752 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-log-socket\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997413 4752 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-etc-openvswitch\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997450 4752 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-kubelet\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:06.998295 master-0 kubenswrapper[4752]: I1205 10:37:06.997469 4752 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-host-run-netns\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:07.001060 master-0 kubenswrapper[4752]: I1205 10:37:07.001018 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2a62edc-8f1c-43be-b355-8484bbebac86-kube-api-access-vnxbz" (OuterVolumeSpecName: "kube-api-access-vnxbz") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "kube-api-access-vnxbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:37:07.002498 master-0 kubenswrapper[4752]: I1205 10:37:07.002457 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2a62edc-8f1c-43be-b355-8484bbebac86-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:37:07.003201 master-0 kubenswrapper[4752]: I1205 10:37:07.003127 4752 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "b2a62edc-8f1c-43be-b355-8484bbebac86" (UID: "b2a62edc-8f1c-43be-b355-8484bbebac86"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:07.098785 master-0 kubenswrapper[4752]: I1205 10:37:07.098629 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.098785 master-0 kubenswrapper[4752]: I1205 10:37:07.098755 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-systemd-units\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099163 master-0 kubenswrapper[4752]: I1205 10:37:07.098884 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-systemd-units\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099163 master-0 kubenswrapper[4752]: I1205 10:37:07.098952 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-netd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099163 master-0 kubenswrapper[4752]: I1205 10:37:07.099009 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-netns\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099163 master-0 kubenswrapper[4752]: I1205 10:37:07.099036 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099163 master-0 kubenswrapper[4752]: I1205 10:37:07.099067 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099163 master-0 kubenswrapper[4752]: I1205 10:37:07.099125 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099163 master-0 kubenswrapper[4752]: I1205 10:37:07.099150 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-bin\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099664 master-0 kubenswrapper[4752]: I1205 10:37:07.099190 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-env-overrides\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099664 master-0 kubenswrapper[4752]: I1205 10:37:07.099213 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-netns\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099664 master-0 kubenswrapper[4752]: I1205 10:37:07.099234 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-node-log\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099664 master-0 kubenswrapper[4752]: I1205 10:37:07.099067 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-netd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099664 master-0 kubenswrapper[4752]: I1205 10:37:07.099470 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-bin\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099664 master-0 kubenswrapper[4752]: I1205 10:37:07.099546 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-systemd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099664 master-0 kubenswrapper[4752]: I1205 10:37:07.099603 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-systemd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.099664 master-0 kubenswrapper[4752]: I1205 10:37:07.099623 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-node-log\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100091 master-0 kubenswrapper[4752]: I1205 10:37:07.099638 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-config\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100091 master-0 kubenswrapper[4752]: I1205 10:37:07.099750 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/495ba1ea-f844-43ec-8be7-47e738f5428a-ovn-node-metrics-cert\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100091 master-0 kubenswrapper[4752]: I1205 10:37:07.099786 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-script-lib\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100091 master-0 kubenswrapper[4752]: I1205 10:37:07.099854 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-etc-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100091 master-0 kubenswrapper[4752]: I1205 10:37:07.099922 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-var-lib-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100091 master-0 kubenswrapper[4752]: I1205 10:37:07.100005 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-var-lib-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100091 master-0 kubenswrapper[4752]: I1205 10:37:07.100070 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-log-socket\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100076 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-etc-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100129 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-log-socket\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100205 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100256 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-ovn\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100308 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100310 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-kubelet\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100340 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-ovn\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100363 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-kubelet\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100356 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjgc4\" (UniqueName: \"kubernetes.io/projected/495ba1ea-f844-43ec-8be7-47e738f5428a-kube-api-access-tjgc4\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100390 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-slash\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100439 4752 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnxbz\" (UniqueName: \"kubernetes.io/projected/b2a62edc-8f1c-43be-b355-8484bbebac86-kube-api-access-vnxbz\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100450 4752 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b2a62edc-8f1c-43be-b355-8484bbebac86-ovn-node-metrics-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100461 4752 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b2a62edc-8f1c-43be-b355-8484bbebac86-run-systemd\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:07.100601 master-0 kubenswrapper[4752]: I1205 10:37:07.100483 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-slash\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.101654 master-0 kubenswrapper[4752]: I1205 10:37:07.100622 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-env-overrides\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.101654 master-0 kubenswrapper[4752]: I1205 10:37:07.100695 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-script-lib\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.101654 master-0 kubenswrapper[4752]: I1205 10:37:07.100739 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-config\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.105452 master-0 kubenswrapper[4752]: I1205 10:37:07.105318 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/495ba1ea-f844-43ec-8be7-47e738f5428a-ovn-node-metrics-cert\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.132119 master-0 kubenswrapper[4752]: I1205 10:37:07.132028 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjgc4\" (UniqueName: \"kubernetes.io/projected/495ba1ea-f844-43ec-8be7-47e738f5428a-kube-api-access-tjgc4\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.207976 master-0 kubenswrapper[4752]: I1205 10:37:07.207788 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d89ht_b2a62edc-8f1c-43be-b355-8484bbebac86/ovnkube-controller/0.log" Dec 05 10:37:07.209286 master-0 kubenswrapper[4752]: I1205 10:37:07.209259 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d89ht_b2a62edc-8f1c-43be-b355-8484bbebac86/kube-rbac-proxy-ovn-metrics/0.log" Dec 05 10:37:07.209755 master-0 kubenswrapper[4752]: I1205 10:37:07.209723 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d89ht_b2a62edc-8f1c-43be-b355-8484bbebac86/kube-rbac-proxy-node/0.log" Dec 05 10:37:07.210095 master-0 kubenswrapper[4752]: I1205 10:37:07.210072 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d89ht_b2a62edc-8f1c-43be-b355-8484bbebac86/ovn-acl-logging/0.log" Dec 05 10:37:07.210475 master-0 kubenswrapper[4752]: I1205 10:37:07.210443 4752 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-d89ht_b2a62edc-8f1c-43be-b355-8484bbebac86/ovn-controller/0.log" Dec 05 10:37:07.210735 master-0 kubenswrapper[4752]: I1205 10:37:07.210700 4752 generic.go:334] "Generic (PLEG): container finished" podID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerID="00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b" exitCode=1 Dec 05 10:37:07.210735 master-0 kubenswrapper[4752]: I1205 10:37:07.210726 4752 generic.go:334] "Generic (PLEG): container finished" podID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerID="573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98" exitCode=0 Dec 05 10:37:07.210735 master-0 kubenswrapper[4752]: I1205 10:37:07.210735 4752 generic.go:334] "Generic (PLEG): container finished" podID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerID="cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87" exitCode=0 Dec 05 10:37:07.210735 master-0 kubenswrapper[4752]: I1205 10:37:07.210742 4752 generic.go:334] "Generic (PLEG): container finished" podID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerID="29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3" exitCode=0 Dec 05 10:37:07.211020 master-0 kubenswrapper[4752]: I1205 10:37:07.210750 4752 generic.go:334] "Generic (PLEG): container finished" podID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerID="88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049" exitCode=143 Dec 05 10:37:07.211020 master-0 kubenswrapper[4752]: I1205 10:37:07.210758 4752 generic.go:334] "Generic (PLEG): container finished" podID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerID="b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5" exitCode=143 Dec 05 10:37:07.211020 master-0 kubenswrapper[4752]: I1205 10:37:07.210765 4752 generic.go:334] "Generic (PLEG): container finished" podID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerID="9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2" exitCode=143 Dec 05 10:37:07.211020 master-0 kubenswrapper[4752]: I1205 10:37:07.210768 4752 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" Dec 05 10:37:07.211020 master-0 kubenswrapper[4752]: I1205 10:37:07.210790 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerDied","Data":"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b"} Dec 05 10:37:07.211020 master-0 kubenswrapper[4752]: I1205 10:37:07.210830 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerDied","Data":"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98"} Dec 05 10:37:07.211020 master-0 kubenswrapper[4752]: I1205 10:37:07.210844 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerDied","Data":"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87"} Dec 05 10:37:07.211020 master-0 kubenswrapper[4752]: I1205 10:37:07.210857 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerDied","Data":"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3"} Dec 05 10:37:07.211020 master-0 kubenswrapper[4752]: I1205 10:37:07.210955 4752 scope.go:117] "RemoveContainer" containerID="00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b" Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211227 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerDied","Data":"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.210774 4752 generic.go:334] "Generic (PLEG): container finished" podID="b2a62edc-8f1c-43be-b355-8484bbebac86" containerID="21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e" exitCode=143 Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211334 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerDied","Data":"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211489 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211589 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211604 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211618 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerDied","Data":"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211636 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211647 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211664 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211671 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211676 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211681 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211687 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211693 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211699 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211709 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerDied","Data":"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211722 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211731 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211738 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87"} Dec 05 10:37:07.211685 master-0 kubenswrapper[4752]: I1205 10:37:07.211744 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211751 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211757 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211767 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211773 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211778 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211786 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-d89ht" event={"ID":"b2a62edc-8f1c-43be-b355-8484bbebac86","Type":"ContainerDied","Data":"a20c01724bf2a194e055334e233e74f0b08a60cb52f191783b538ff245ddd1cb"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211797 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211805 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211811 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211817 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211825 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211831 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211836 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211842 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e"} Dec 05 10:37:07.214269 master-0 kubenswrapper[4752]: I1205 10:37:07.211848 4752 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f"} Dec 05 10:37:07.236682 master-0 kubenswrapper[4752]: I1205 10:37:07.236629 4752 scope.go:117] "RemoveContainer" containerID="573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98" Dec 05 10:37:07.249568 master-0 kubenswrapper[4752]: I1205 10:37:07.249162 4752 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-d89ht"] Dec 05 10:37:07.250261 master-0 kubenswrapper[4752]: I1205 10:37:07.250067 4752 scope.go:117] "RemoveContainer" containerID="cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87" Dec 05 10:37:07.252826 master-0 kubenswrapper[4752]: I1205 10:37:07.252657 4752 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-d89ht"] Dec 05 10:37:07.263945 master-0 kubenswrapper[4752]: I1205 10:37:07.263830 4752 scope.go:117] "RemoveContainer" containerID="29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3" Dec 05 10:37:07.273097 master-0 kubenswrapper[4752]: I1205 10:37:07.273030 4752 scope.go:117] "RemoveContainer" containerID="88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049" Dec 05 10:37:07.276263 master-0 kubenswrapper[4752]: I1205 10:37:07.276217 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:07.285908 master-0 kubenswrapper[4752]: I1205 10:37:07.285865 4752 scope.go:117] "RemoveContainer" containerID="b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5" Dec 05 10:37:07.290825 master-0 kubenswrapper[4752]: W1205 10:37:07.290769 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod495ba1ea_f844_43ec_8be7_47e738f5428a.slice/crio-fc75409854b1370368a39cd91bcd5c4ac2aa319c2a8b9aaad1e7abab3fd56a49 WatchSource:0}: Error finding container fc75409854b1370368a39cd91bcd5c4ac2aa319c2a8b9aaad1e7abab3fd56a49: Status 404 returned error can't find the container with id fc75409854b1370368a39cd91bcd5c4ac2aa319c2a8b9aaad1e7abab3fd56a49 Dec 05 10:37:07.294674 master-0 kubenswrapper[4752]: I1205 10:37:07.294626 4752 scope.go:117] "RemoveContainer" containerID="9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2" Dec 05 10:37:07.303395 master-0 kubenswrapper[4752]: I1205 10:37:07.303305 4752 scope.go:117] "RemoveContainer" containerID="21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e" Dec 05 10:37:07.318301 master-0 kubenswrapper[4752]: I1205 10:37:07.318228 4752 scope.go:117] "RemoveContainer" containerID="3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f" Dec 05 10:37:07.327166 master-0 kubenswrapper[4752]: I1205 10:37:07.326753 4752 scope.go:117] "RemoveContainer" containerID="00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b" Dec 05 10:37:07.327676 master-0 kubenswrapper[4752]: E1205 10:37:07.327632 4752 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b\": container with ID starting with 00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b not found: ID does not exist" containerID="00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b" Dec 05 10:37:07.327735 master-0 kubenswrapper[4752]: I1205 10:37:07.327681 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b"} err="failed to get container status \"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b\": rpc error: code = NotFound desc = could not find container \"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b\": container with ID starting with 00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b not found: ID does not exist" Dec 05 10:37:07.327735 master-0 kubenswrapper[4752]: I1205 10:37:07.327718 4752 scope.go:117] "RemoveContainer" containerID="573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98" Dec 05 10:37:07.328065 master-0 kubenswrapper[4752]: E1205 10:37:07.328020 4752 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98\": container with ID starting with 573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98 not found: ID does not exist" containerID="573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98" Dec 05 10:37:07.328145 master-0 kubenswrapper[4752]: I1205 10:37:07.328108 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98"} err="failed to get container status \"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98\": rpc error: code = NotFound desc = could not find container \"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98\": container with ID starting with 573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98 not found: ID does not exist" Dec 05 10:37:07.328188 master-0 kubenswrapper[4752]: I1205 10:37:07.328146 4752 scope.go:117] "RemoveContainer" containerID="cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87" Dec 05 10:37:07.328518 master-0 kubenswrapper[4752]: E1205 10:37:07.328485 4752 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87\": container with ID starting with cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87 not found: ID does not exist" containerID="cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87" Dec 05 10:37:07.328587 master-0 kubenswrapper[4752]: I1205 10:37:07.328512 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87"} err="failed to get container status \"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87\": rpc error: code = NotFound desc = could not find container \"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87\": container with ID starting with cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87 not found: ID does not exist" Dec 05 10:37:07.328587 master-0 kubenswrapper[4752]: I1205 10:37:07.328533 4752 scope.go:117] "RemoveContainer" containerID="29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3" Dec 05 10:37:07.328846 master-0 kubenswrapper[4752]: E1205 10:37:07.328815 4752 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3\": container with ID starting with 29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3 not found: ID does not exist" containerID="29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3" Dec 05 10:37:07.328846 master-0 kubenswrapper[4752]: I1205 10:37:07.328838 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3"} err="failed to get container status \"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3\": rpc error: code = NotFound desc = could not find container \"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3\": container with ID starting with 29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3 not found: ID does not exist" Dec 05 10:37:07.328937 master-0 kubenswrapper[4752]: I1205 10:37:07.328855 4752 scope.go:117] "RemoveContainer" containerID="88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049" Dec 05 10:37:07.329130 master-0 kubenswrapper[4752]: E1205 10:37:07.329100 4752 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049\": container with ID starting with 88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049 not found: ID does not exist" containerID="88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049" Dec 05 10:37:07.329130 master-0 kubenswrapper[4752]: I1205 10:37:07.329121 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049"} err="failed to get container status \"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049\": rpc error: code = NotFound desc = could not find container \"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049\": container with ID starting with 88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049 not found: ID does not exist" Dec 05 10:37:07.329233 master-0 kubenswrapper[4752]: I1205 10:37:07.329137 4752 scope.go:117] "RemoveContainer" containerID="b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5" Dec 05 10:37:07.329635 master-0 kubenswrapper[4752]: E1205 10:37:07.329587 4752 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5\": container with ID starting with b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5 not found: ID does not exist" containerID="b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5" Dec 05 10:37:07.329699 master-0 kubenswrapper[4752]: I1205 10:37:07.329642 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5"} err="failed to get container status \"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5\": rpc error: code = NotFound desc = could not find container \"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5\": container with ID starting with b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5 not found: ID does not exist" Dec 05 10:37:07.329699 master-0 kubenswrapper[4752]: I1205 10:37:07.329683 4752 scope.go:117] "RemoveContainer" containerID="9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2" Dec 05 10:37:07.329990 master-0 kubenswrapper[4752]: E1205 10:37:07.329961 4752 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2\": container with ID starting with 9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2 not found: ID does not exist" containerID="9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2" Dec 05 10:37:07.329990 master-0 kubenswrapper[4752]: I1205 10:37:07.329982 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2"} err="failed to get container status \"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2\": rpc error: code = NotFound desc = could not find container \"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2\": container with ID starting with 9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2 not found: ID does not exist" Dec 05 10:37:07.330088 master-0 kubenswrapper[4752]: I1205 10:37:07.329999 4752 scope.go:117] "RemoveContainer" containerID="21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e" Dec 05 10:37:07.330290 master-0 kubenswrapper[4752]: E1205 10:37:07.330246 4752 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e\": container with ID starting with 21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e not found: ID does not exist" containerID="21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e" Dec 05 10:37:07.330348 master-0 kubenswrapper[4752]: I1205 10:37:07.330287 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e"} err="failed to get container status \"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e\": rpc error: code = NotFound desc = could not find container \"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e\": container with ID starting with 21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e not found: ID does not exist" Dec 05 10:37:07.330348 master-0 kubenswrapper[4752]: I1205 10:37:07.330315 4752 scope.go:117] "RemoveContainer" containerID="3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f" Dec 05 10:37:07.330722 master-0 kubenswrapper[4752]: E1205 10:37:07.330689 4752 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f\": container with ID starting with 3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f not found: ID does not exist" containerID="3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f" Dec 05 10:37:07.330722 master-0 kubenswrapper[4752]: I1205 10:37:07.330715 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f"} err="failed to get container status \"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f\": rpc error: code = NotFound desc = could not find container \"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f\": container with ID starting with 3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f not found: ID does not exist" Dec 05 10:37:07.330829 master-0 kubenswrapper[4752]: I1205 10:37:07.330732 4752 scope.go:117] "RemoveContainer" containerID="00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b" Dec 05 10:37:07.331140 master-0 kubenswrapper[4752]: I1205 10:37:07.331102 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b"} err="failed to get container status \"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b\": rpc error: code = NotFound desc = could not find container \"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b\": container with ID starting with 00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b not found: ID does not exist" Dec 05 10:37:07.331140 master-0 kubenswrapper[4752]: I1205 10:37:07.331127 4752 scope.go:117] "RemoveContainer" containerID="573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98" Dec 05 10:37:07.331430 master-0 kubenswrapper[4752]: I1205 10:37:07.331382 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98"} err="failed to get container status \"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98\": rpc error: code = NotFound desc = could not find container \"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98\": container with ID starting with 573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98 not found: ID does not exist" Dec 05 10:37:07.331430 master-0 kubenswrapper[4752]: I1205 10:37:07.331401 4752 scope.go:117] "RemoveContainer" containerID="cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87" Dec 05 10:37:07.331797 master-0 kubenswrapper[4752]: I1205 10:37:07.331763 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87"} err="failed to get container status \"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87\": rpc error: code = NotFound desc = could not find container \"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87\": container with ID starting with cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87 not found: ID does not exist" Dec 05 10:37:07.331797 master-0 kubenswrapper[4752]: I1205 10:37:07.331785 4752 scope.go:117] "RemoveContainer" containerID="29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3" Dec 05 10:37:07.332118 master-0 kubenswrapper[4752]: I1205 10:37:07.332066 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3"} err="failed to get container status \"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3\": rpc error: code = NotFound desc = could not find container \"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3\": container with ID starting with 29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3 not found: ID does not exist" Dec 05 10:37:07.332180 master-0 kubenswrapper[4752]: I1205 10:37:07.332117 4752 scope.go:117] "RemoveContainer" containerID="88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049" Dec 05 10:37:07.332486 master-0 kubenswrapper[4752]: I1205 10:37:07.332443 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049"} err="failed to get container status \"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049\": rpc error: code = NotFound desc = could not find container \"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049\": container with ID starting with 88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049 not found: ID does not exist" Dec 05 10:37:07.332486 master-0 kubenswrapper[4752]: I1205 10:37:07.332472 4752 scope.go:117] "RemoveContainer" containerID="b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5" Dec 05 10:37:07.332961 master-0 kubenswrapper[4752]: I1205 10:37:07.332917 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5"} err="failed to get container status \"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5\": rpc error: code = NotFound desc = could not find container \"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5\": container with ID starting with b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5 not found: ID does not exist" Dec 05 10:37:07.332961 master-0 kubenswrapper[4752]: I1205 10:37:07.332953 4752 scope.go:117] "RemoveContainer" containerID="9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2" Dec 05 10:37:07.333250 master-0 kubenswrapper[4752]: I1205 10:37:07.333218 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2"} err="failed to get container status \"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2\": rpc error: code = NotFound desc = could not find container \"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2\": container with ID starting with 9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2 not found: ID does not exist" Dec 05 10:37:07.333250 master-0 kubenswrapper[4752]: I1205 10:37:07.333239 4752 scope.go:117] "RemoveContainer" containerID="21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e" Dec 05 10:37:07.333569 master-0 kubenswrapper[4752]: I1205 10:37:07.333532 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e"} err="failed to get container status \"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e\": rpc error: code = NotFound desc = could not find container \"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e\": container with ID starting with 21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e not found: ID does not exist" Dec 05 10:37:07.333569 master-0 kubenswrapper[4752]: I1205 10:37:07.333555 4752 scope.go:117] "RemoveContainer" containerID="3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f" Dec 05 10:37:07.333802 master-0 kubenswrapper[4752]: I1205 10:37:07.333771 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f"} err="failed to get container status \"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f\": rpc error: code = NotFound desc = could not find container \"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f\": container with ID starting with 3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f not found: ID does not exist" Dec 05 10:37:07.333802 master-0 kubenswrapper[4752]: I1205 10:37:07.333791 4752 scope.go:117] "RemoveContainer" containerID="00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b" Dec 05 10:37:07.334147 master-0 kubenswrapper[4752]: I1205 10:37:07.334109 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b"} err="failed to get container status \"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b\": rpc error: code = NotFound desc = could not find container \"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b\": container with ID starting with 00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b not found: ID does not exist" Dec 05 10:37:07.334147 master-0 kubenswrapper[4752]: I1205 10:37:07.334131 4752 scope.go:117] "RemoveContainer" containerID="573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98" Dec 05 10:37:07.334393 master-0 kubenswrapper[4752]: I1205 10:37:07.334362 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98"} err="failed to get container status \"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98\": rpc error: code = NotFound desc = could not find container \"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98\": container with ID starting with 573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98 not found: ID does not exist" Dec 05 10:37:07.334393 master-0 kubenswrapper[4752]: I1205 10:37:07.334384 4752 scope.go:117] "RemoveContainer" containerID="cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87" Dec 05 10:37:07.334645 master-0 kubenswrapper[4752]: I1205 10:37:07.334613 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87"} err="failed to get container status \"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87\": rpc error: code = NotFound desc = could not find container \"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87\": container with ID starting with cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87 not found: ID does not exist" Dec 05 10:37:07.334645 master-0 kubenswrapper[4752]: I1205 10:37:07.334635 4752 scope.go:117] "RemoveContainer" containerID="29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3" Dec 05 10:37:07.334900 master-0 kubenswrapper[4752]: I1205 10:37:07.334865 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3"} err="failed to get container status \"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3\": rpc error: code = NotFound desc = could not find container \"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3\": container with ID starting with 29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3 not found: ID does not exist" Dec 05 10:37:07.334900 master-0 kubenswrapper[4752]: I1205 10:37:07.334885 4752 scope.go:117] "RemoveContainer" containerID="88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049" Dec 05 10:37:07.335117 master-0 kubenswrapper[4752]: I1205 10:37:07.335086 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049"} err="failed to get container status \"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049\": rpc error: code = NotFound desc = could not find container \"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049\": container with ID starting with 88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049 not found: ID does not exist" Dec 05 10:37:07.335117 master-0 kubenswrapper[4752]: I1205 10:37:07.335109 4752 scope.go:117] "RemoveContainer" containerID="b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5" Dec 05 10:37:07.335497 master-0 kubenswrapper[4752]: I1205 10:37:07.335461 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5"} err="failed to get container status \"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5\": rpc error: code = NotFound desc = could not find container \"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5\": container with ID starting with b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5 not found: ID does not exist" Dec 05 10:37:07.335497 master-0 kubenswrapper[4752]: I1205 10:37:07.335486 4752 scope.go:117] "RemoveContainer" containerID="9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2" Dec 05 10:37:07.335831 master-0 kubenswrapper[4752]: I1205 10:37:07.335784 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2"} err="failed to get container status \"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2\": rpc error: code = NotFound desc = could not find container \"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2\": container with ID starting with 9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2 not found: ID does not exist" Dec 05 10:37:07.335831 master-0 kubenswrapper[4752]: I1205 10:37:07.335815 4752 scope.go:117] "RemoveContainer" containerID="21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e" Dec 05 10:37:07.336063 master-0 kubenswrapper[4752]: I1205 10:37:07.336026 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e"} err="failed to get container status \"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e\": rpc error: code = NotFound desc = could not find container \"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e\": container with ID starting with 21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e not found: ID does not exist" Dec 05 10:37:07.336063 master-0 kubenswrapper[4752]: I1205 10:37:07.336050 4752 scope.go:117] "RemoveContainer" containerID="3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f" Dec 05 10:37:07.336344 master-0 kubenswrapper[4752]: I1205 10:37:07.336301 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f"} err="failed to get container status \"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f\": rpc error: code = NotFound desc = could not find container \"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f\": container with ID starting with 3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f not found: ID does not exist" Dec 05 10:37:07.336401 master-0 kubenswrapper[4752]: I1205 10:37:07.336341 4752 scope.go:117] "RemoveContainer" containerID="00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b" Dec 05 10:37:07.336692 master-0 kubenswrapper[4752]: I1205 10:37:07.336662 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b"} err="failed to get container status \"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b\": rpc error: code = NotFound desc = could not find container \"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b\": container with ID starting with 00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b not found: ID does not exist" Dec 05 10:37:07.336692 master-0 kubenswrapper[4752]: I1205 10:37:07.336680 4752 scope.go:117] "RemoveContainer" containerID="573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98" Dec 05 10:37:07.336950 master-0 kubenswrapper[4752]: I1205 10:37:07.336920 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98"} err="failed to get container status \"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98\": rpc error: code = NotFound desc = could not find container \"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98\": container with ID starting with 573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98 not found: ID does not exist" Dec 05 10:37:07.336950 master-0 kubenswrapper[4752]: I1205 10:37:07.336939 4752 scope.go:117] "RemoveContainer" containerID="cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87" Dec 05 10:37:07.337285 master-0 kubenswrapper[4752]: I1205 10:37:07.337225 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87"} err="failed to get container status \"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87\": rpc error: code = NotFound desc = could not find container \"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87\": container with ID starting with cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87 not found: ID does not exist" Dec 05 10:37:07.337336 master-0 kubenswrapper[4752]: I1205 10:37:07.337282 4752 scope.go:117] "RemoveContainer" containerID="29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3" Dec 05 10:37:07.337576 master-0 kubenswrapper[4752]: I1205 10:37:07.337545 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3"} err="failed to get container status \"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3\": rpc error: code = NotFound desc = could not find container \"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3\": container with ID starting with 29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3 not found: ID does not exist" Dec 05 10:37:07.337576 master-0 kubenswrapper[4752]: I1205 10:37:07.337564 4752 scope.go:117] "RemoveContainer" containerID="88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049" Dec 05 10:37:07.337831 master-0 kubenswrapper[4752]: I1205 10:37:07.337787 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049"} err="failed to get container status \"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049\": rpc error: code = NotFound desc = could not find container \"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049\": container with ID starting with 88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049 not found: ID does not exist" Dec 05 10:37:07.337831 master-0 kubenswrapper[4752]: I1205 10:37:07.337824 4752 scope.go:117] "RemoveContainer" containerID="b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5" Dec 05 10:37:07.338076 master-0 kubenswrapper[4752]: I1205 10:37:07.338045 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5"} err="failed to get container status \"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5\": rpc error: code = NotFound desc = could not find container \"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5\": container with ID starting with b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5 not found: ID does not exist" Dec 05 10:37:07.338076 master-0 kubenswrapper[4752]: I1205 10:37:07.338064 4752 scope.go:117] "RemoveContainer" containerID="9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2" Dec 05 10:37:07.338320 master-0 kubenswrapper[4752]: I1205 10:37:07.338280 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2"} err="failed to get container status \"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2\": rpc error: code = NotFound desc = could not find container \"9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2\": container with ID starting with 9b87e961e41f23aeac7b037fccf51d0819abeb8de4e4fb388dc2511ff61394b2 not found: ID does not exist" Dec 05 10:37:07.338320 master-0 kubenswrapper[4752]: I1205 10:37:07.338314 4752 scope.go:117] "RemoveContainer" containerID="21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e" Dec 05 10:37:07.338564 master-0 kubenswrapper[4752]: I1205 10:37:07.338539 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e"} err="failed to get container status \"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e\": rpc error: code = NotFound desc = could not find container \"21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e\": container with ID starting with 21ef1131166e540e0cf10c94a59ee74949b32a6fccba40f40f010f87ca828e9e not found: ID does not exist" Dec 05 10:37:07.338564 master-0 kubenswrapper[4752]: I1205 10:37:07.338561 4752 scope.go:117] "RemoveContainer" containerID="3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f" Dec 05 10:37:07.338787 master-0 kubenswrapper[4752]: I1205 10:37:07.338755 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f"} err="failed to get container status \"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f\": rpc error: code = NotFound desc = could not find container \"3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f\": container with ID starting with 3d844ca6a2d89dc99b07d7bb0e29d6b539380d31fcdae4f7fff0cd387006460f not found: ID does not exist" Dec 05 10:37:07.338835 master-0 kubenswrapper[4752]: I1205 10:37:07.338788 4752 scope.go:117] "RemoveContainer" containerID="00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b" Dec 05 10:37:07.339022 master-0 kubenswrapper[4752]: I1205 10:37:07.338998 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b"} err="failed to get container status \"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b\": rpc error: code = NotFound desc = could not find container \"00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b\": container with ID starting with 00644dcaf45e311174a4190c805a26b067405299ea4f1e9bda43e108f2f5007b not found: ID does not exist" Dec 05 10:37:07.339022 master-0 kubenswrapper[4752]: I1205 10:37:07.339019 4752 scope.go:117] "RemoveContainer" containerID="573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98" Dec 05 10:37:07.339334 master-0 kubenswrapper[4752]: I1205 10:37:07.339292 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98"} err="failed to get container status \"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98\": rpc error: code = NotFound desc = could not find container \"573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98\": container with ID starting with 573d0e0b1a9001e14ebd40c36684b893ebb8eae968ece22e5f35060091fbea98 not found: ID does not exist" Dec 05 10:37:07.339394 master-0 kubenswrapper[4752]: I1205 10:37:07.339335 4752 scope.go:117] "RemoveContainer" containerID="cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87" Dec 05 10:37:07.339703 master-0 kubenswrapper[4752]: I1205 10:37:07.339663 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87"} err="failed to get container status \"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87\": rpc error: code = NotFound desc = could not find container \"cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87\": container with ID starting with cc05edea1496a464d6fbdf920e82eec31a1f452d35479e3b845afc2930f80a87 not found: ID does not exist" Dec 05 10:37:07.339703 master-0 kubenswrapper[4752]: I1205 10:37:07.339698 4752 scope.go:117] "RemoveContainer" containerID="29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3" Dec 05 10:37:07.339968 master-0 kubenswrapper[4752]: I1205 10:37:07.339937 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3"} err="failed to get container status \"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3\": rpc error: code = NotFound desc = could not find container \"29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3\": container with ID starting with 29861ebbcdbea40c2037956b4817dca358e81489dd4caff3d3af2faed3d70ab3 not found: ID does not exist" Dec 05 10:37:07.339968 master-0 kubenswrapper[4752]: I1205 10:37:07.339956 4752 scope.go:117] "RemoveContainer" containerID="88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049" Dec 05 10:37:07.340333 master-0 kubenswrapper[4752]: I1205 10:37:07.340236 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049"} err="failed to get container status \"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049\": rpc error: code = NotFound desc = could not find container \"88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049\": container with ID starting with 88ce96ba5a7adc6948881773fbc6d5c31fe87e519393e306bdda7c584ae83049 not found: ID does not exist" Dec 05 10:37:07.340333 master-0 kubenswrapper[4752]: I1205 10:37:07.340281 4752 scope.go:117] "RemoveContainer" containerID="b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5" Dec 05 10:37:07.340684 master-0 kubenswrapper[4752]: I1205 10:37:07.340632 4752 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5"} err="failed to get container status \"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5\": rpc error: code = NotFound desc = could not find container \"b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5\": container with ID starting with b4e2cee8cb44d23935a233d893e957190e5e5194c0a5f331b052874ccde452e5 not found: ID does not exist" Dec 05 10:37:07.796017 master-0 kubenswrapper[4752]: I1205 10:37:07.795962 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:07.796017 master-0 kubenswrapper[4752]: I1205 10:37:07.796020 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:07.796269 master-0 kubenswrapper[4752]: E1205 10:37:07.796158 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8gjgm" podUID="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" Dec 05 10:37:07.796313 master-0 kubenswrapper[4752]: E1205 10:37:07.796263 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-d6fzk" podUID="720a7467-ce93-4d48-82ec-9ad0922d99c2" Dec 05 10:37:08.216284 master-0 kubenswrapper[4752]: I1205 10:37:08.216219 4752 generic.go:334] "Generic (PLEG): container finished" podID="495ba1ea-f844-43ec-8be7-47e738f5428a" containerID="f02b670dd80cfe6eb6e6f824971df1d4d23971943a981cfcb563de519950bb3e" exitCode=0 Dec 05 10:37:08.216284 master-0 kubenswrapper[4752]: I1205 10:37:08.216277 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" event={"ID":"495ba1ea-f844-43ec-8be7-47e738f5428a","Type":"ContainerDied","Data":"f02b670dd80cfe6eb6e6f824971df1d4d23971943a981cfcb563de519950bb3e"} Dec 05 10:37:08.217272 master-0 kubenswrapper[4752]: I1205 10:37:08.216323 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" event={"ID":"495ba1ea-f844-43ec-8be7-47e738f5428a","Type":"ContainerStarted","Data":"fc75409854b1370368a39cd91bcd5c4ac2aa319c2a8b9aaad1e7abab3fd56a49"} Dec 05 10:37:08.426290 master-0 kubenswrapper[4752]: I1205 10:37:08.426260 4752 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeReady" Dec 05 10:37:08.426454 master-0 kubenswrapper[4752]: I1205 10:37:08.426435 4752 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Dec 05 10:37:08.455310 master-0 kubenswrapper[4752]: I1205 10:37:08.455271 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr"] Dec 05 10:37:08.455876 master-0 kubenswrapper[4752]: I1205 10:37:08.455856 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:08.458091 master-0 kubenswrapper[4752]: I1205 10:37:08.458068 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 10:37:08.458482 master-0 kubenswrapper[4752]: I1205 10:37:08.458398 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.458552 master-0 kubenswrapper[4752]: I1205 10:37:08.458295 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 10:37:08.458622 master-0 kubenswrapper[4752]: I1205 10:37:08.458332 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 10:37:08.460081 master-0 kubenswrapper[4752]: I1205 10:37:08.459667 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p"] Dec 05 10:37:08.460572 master-0 kubenswrapper[4752]: I1205 10:37:08.460250 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27"] Dec 05 10:37:08.461397 master-0 kubenswrapper[4752]: I1205 10:37:08.460996 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk"] Dec 05 10:37:08.461397 master-0 kubenswrapper[4752]: I1205 10:37:08.461360 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8"] Dec 05 10:37:08.461816 master-0 kubenswrapper[4752]: I1205 10:37:08.461771 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c"] Dec 05 10:37:08.462322 master-0 kubenswrapper[4752]: I1205 10:37:08.462265 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f"] Dec 05 10:37:08.462385 master-0 kubenswrapper[4752]: I1205 10:37:08.462348 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:08.462618 master-0 kubenswrapper[4752]: I1205 10:37:08.462274 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.462738 master-0 kubenswrapper[4752]: I1205 10:37:08.462716 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:08.463043 master-0 kubenswrapper[4752]: I1205 10:37:08.462984 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:08.464509 master-0 kubenswrapper[4752]: I1205 10:37:08.464490 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7"] Dec 05 10:37:08.465524 master-0 kubenswrapper[4752]: I1205 10:37:08.465446 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-f797b99b6-z9qcl"] Dec 05 10:37:08.465774 master-0 kubenswrapper[4752]: I1205 10:37:08.464881 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.465836 master-0 kubenswrapper[4752]: I1205 10:37:08.465151 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.465887 master-0 kubenswrapper[4752]: I1205 10:37:08.464852 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:08.465956 master-0 kubenswrapper[4752]: I1205 10:37:08.465755 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4"] Dec 05 10:37:08.467444 master-0 kubenswrapper[4752]: I1205 10:37:08.465399 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.467444 master-0 kubenswrapper[4752]: I1205 10:37:08.467291 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-storage-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.467444 master-0 kubenswrapper[4752]: I1205 10:37:08.466039 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:08.467687 master-0 kubenswrapper[4752]: I1205 10:37:08.467666 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 10:37:08.467824 master-0 kubenswrapper[4752]: I1205 10:37:08.467796 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:08.467875 master-0 kubenswrapper[4752]: I1205 10:37:08.467846 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 10:37:08.468056 master-0 kubenswrapper[4752]: I1205 10:37:08.468016 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn"] Dec 05 10:37:08.468380 master-0 kubenswrapper[4752]: I1205 10:37:08.468354 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:08.468481 master-0 kubenswrapper[4752]: I1205 10:37:08.468079 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 10:37:08.468576 master-0 kubenswrapper[4752]: I1205 10:37:08.468114 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-olm-operator"/"cluster-olm-operator-serving-cert" Dec 05 10:37:08.468606 master-0 kubenswrapper[4752]: I1205 10:37:08.468293 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 10:37:08.468647 master-0 kubenswrapper[4752]: I1205 10:37:08.468363 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5"] Dec 05 10:37:08.468867 master-0 kubenswrapper[4752]: I1205 10:37:08.468853 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr"] Dec 05 10:37:08.469071 master-0 kubenswrapper[4752]: I1205 10:37:08.469058 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl"] Dec 05 10:37:08.469241 master-0 kubenswrapper[4752]: I1205 10:37:08.469217 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:08.469295 master-0 kubenswrapper[4752]: I1205 10:37:08.469256 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:08.469412 master-0 kubenswrapper[4752]: I1205 10:37:08.469219 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:08.469751 master-0 kubenswrapper[4752]: I1205 10:37:08.469728 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-8649c48786-cgt5x"] Dec 05 10:37:08.470034 master-0 kubenswrapper[4752]: I1205 10:37:08.470012 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.470254 master-0 kubenswrapper[4752]: I1205 10:37:08.470239 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h"] Dec 05 10:37:08.470666 master-0 kubenswrapper[4752]: I1205 10:37:08.470649 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb"] Dec 05 10:37:08.470932 master-0 kubenswrapper[4752]: I1205 10:37:08.470917 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.471184 master-0 kubenswrapper[4752]: I1205 10:37:08.470282 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.471485 master-0 kubenswrapper[4752]: I1205 10:37:08.471446 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8"] Dec 05 10:37:08.471583 master-0 kubenswrapper[4752]: I1205 10:37:08.471567 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:08.471829 master-0 kubenswrapper[4752]: I1205 10:37:08.471803 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.474524 master-0 kubenswrapper[4752]: I1205 10:37:08.473613 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t"] Dec 05 10:37:08.474524 master-0 kubenswrapper[4752]: I1205 10:37:08.474115 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.474764 master-0 kubenswrapper[4752]: I1205 10:37:08.474717 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-node-tuning-operator"/"node-tuning-operator-tls" Dec 05 10:37:08.474919 master-0 kubenswrapper[4752]: I1205 10:37:08.474889 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-node-tuning-operator"/"performance-addon-operator-webhook-cert" Dec 05 10:37:08.474919 master-0 kubenswrapper[4752]: I1205 10:37:08.474911 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 10:37:08.475040 master-0 kubenswrapper[4752]: I1205 10:37:08.475020 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-olm-operator"/"kube-root-ca.crt" Dec 05 10:37:08.475110 master-0 kubenswrapper[4752]: I1205 10:37:08.475028 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"kube-root-ca.crt" Dec 05 10:37:08.475201 master-0 kubenswrapper[4752]: I1205 10:37:08.475161 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 10:37:08.475230 master-0 kubenswrapper[4752]: I1205 10:37:08.475208 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-olm-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.475449 master-0 kubenswrapper[4752]: I1205 10:37:08.475399 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.475529 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-storage-operator"/"kube-root-ca.crt" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.475632 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.475697 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.475763 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.475876 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv"] Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.476166 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.476259 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.476313 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.476404 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.476756 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.476808 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.476854 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.476970 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.477125 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.476758 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.476332 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-7dfc5b745f-67rx7"] Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.477246 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.477351 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.477365 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.477708 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.477866 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 10:37:08.478535 master-0 kubenswrapper[4752]: I1205 10:37:08.478033 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:08.479579 master-0 kubenswrapper[4752]: I1205 10:37:08.479523 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 10:37:08.481769 master-0 kubenswrapper[4752]: I1205 10:37:08.481724 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/iptables-alerter-d6wjk"] Dec 05 10:37:08.482999 master-0 kubenswrapper[4752]: I1205 10:37:08.482959 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 10:37:08.483117 master-0 kubenswrapper[4752]: I1205 10:37:08.483074 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:08.486666 master-0 kubenswrapper[4752]: I1205 10:37:08.483440 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 10:37:08.486666 master-0 kubenswrapper[4752]: I1205 10:37:08.483869 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:37:08.486666 master-0 kubenswrapper[4752]: I1205 10:37:08.484225 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 10:37:08.486666 master-0 kubenswrapper[4752]: I1205 10:37:08.485475 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 10:37:08.486666 master-0 kubenswrapper[4752]: I1205 10:37:08.485631 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.487992 master-0 kubenswrapper[4752]: I1205 10:37:08.487952 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Dec 05 10:37:08.494947 master-0 kubenswrapper[4752]: I1205 10:37:08.494838 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Dec 05 10:37:08.495521 master-0 kubenswrapper[4752]: I1205 10:37:08.495443 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Dec 05 10:37:08.495521 master-0 kubenswrapper[4752]: I1205 10:37:08.495478 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Dec 05 10:37:08.495661 master-0 kubenswrapper[4752]: I1205 10:37:08.495559 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 10:37:08.495661 master-0 kubenswrapper[4752]: I1205 10:37:08.495608 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.496019 master-0 kubenswrapper[4752]: I1205 10:37:08.495888 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 10:37:08.496019 master-0 kubenswrapper[4752]: I1205 10:37:08.495913 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 10:37:08.496228 master-0 kubenswrapper[4752]: I1205 10:37:08.496036 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.496228 master-0 kubenswrapper[4752]: I1205 10:37:08.496100 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 10:37:08.496415 master-0 kubenswrapper[4752]: I1205 10:37:08.496387 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 10:37:08.497610 master-0 kubenswrapper[4752]: I1205 10:37:08.497044 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 10:37:08.499161 master-0 kubenswrapper[4752]: I1205 10:37:08.498926 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 10:37:08.499161 master-0 kubenswrapper[4752]: I1205 10:37:08.499053 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 10:37:08.499266 master-0 kubenswrapper[4752]: I1205 10:37:08.499180 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 10:37:08.499914 master-0 kubenswrapper[4752]: I1205 10:37:08.499883 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 10:37:08.500129 master-0 kubenswrapper[4752]: I1205 10:37:08.499993 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 10:37:08.500312 master-0 kubenswrapper[4752]: I1205 10:37:08.500280 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 10:37:08.500392 master-0 kubenswrapper[4752]: I1205 10:37:08.500324 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 10:37:08.500450 master-0 kubenswrapper[4752]: I1205 10:37:08.500433 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 10:37:08.500498 master-0 kubenswrapper[4752]: I1205 10:37:08.500478 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:37:08.500607 master-0 kubenswrapper[4752]: I1205 10:37:08.500558 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 10:37:08.500607 master-0 kubenswrapper[4752]: I1205 10:37:08.500597 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 10:37:08.500676 master-0 kubenswrapper[4752]: I1205 10:37:08.500659 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 10:37:08.501559 master-0 kubenswrapper[4752]: I1205 10:37:08.501501 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 10:37:08.503944 master-0 kubenswrapper[4752]: I1205 10:37:08.503899 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"trusted-ca" Dec 05 10:37:08.508920 master-0 kubenswrapper[4752]: I1205 10:37:08.508873 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 10:37:08.509508 master-0 kubenswrapper[4752]: I1205 10:37:08.509473 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 10:37:08.510364 master-0 kubenswrapper[4752]: I1205 10:37:08.510316 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 10:37:08.510592 master-0 kubenswrapper[4752]: I1205 10:37:08.510555 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 10:37:08.513391 master-0 kubenswrapper[4752]: I1205 10:37:08.513350 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.513476 master-0 kubenswrapper[4752]: I1205 10:37:08.513396 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-trusted-ca\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:08.513524 master-0 kubenswrapper[4752]: I1205 10:37:08.513487 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:08.513571 master-0 kubenswrapper[4752]: I1205 10:37:08.513554 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fbss\" (UniqueName: \"kubernetes.io/projected/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-kube-api-access-2fbss\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.513621 master-0 kubenswrapper[4752]: I1205 10:37:08.513580 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khfxz\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-kube-api-access-khfxz\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.513659 master-0 kubenswrapper[4752]: I1205 10:37:08.513625 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/8c649a16-c187-412e-b5da-62a00bee38ab-kube-api-access-d8pv2\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:08.513709 master-0 kubenswrapper[4752]: I1205 10:37:08.513648 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhjhk\" (UniqueName: \"kubernetes.io/projected/283122ba-be1c-4516-bd0f-df41c13c098b-kube-api-access-vhjhk\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:08.513761 master-0 kubenswrapper[4752]: I1205 10:37:08.513710 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lwgq\" (UniqueName: \"kubernetes.io/projected/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-kube-api-access-2lwgq\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:08.513799 master-0 kubenswrapper[4752]: I1205 10:37:08.513776 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:08.513845 master-0 kubenswrapper[4752]: I1205 10:37:08.513804 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn7rj\" (UniqueName: \"kubernetes.io/projected/eb290494-a456-4f0e-9afc-f20abab1a1bf-kube-api-access-qn7rj\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:08.513885 master-0 kubenswrapper[4752]: I1205 10:37:08.513852 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzwfq\" (UniqueName: \"kubernetes.io/projected/6f76d12f-5406-47e2-8337-2f50e35376d6-kube-api-access-fzwfq\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:08.513940 master-0 kubenswrapper[4752]: I1205 10:37:08.513878 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkqz7\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-kube-api-access-vkqz7\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.513994 master-0 kubenswrapper[4752]: I1205 10:37:08.513946 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-serving-cert\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.514037 master-0 kubenswrapper[4752]: I1205 10:37:08.514010 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-service-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.514037 master-0 kubenswrapper[4752]: I1205 10:37:08.514019 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 10:37:08.514113 master-0 kubenswrapper[4752]: I1205 10:37:08.514040 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:08.514254 master-0 kubenswrapper[4752]: I1205 10:37:08.514143 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 10:37:08.514368 master-0 kubenswrapper[4752]: I1205 10:37:08.514242 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2vd4\" (UniqueName: \"kubernetes.io/projected/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-kube-api-access-h2vd4\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:08.514468 master-0 kubenswrapper[4752]: I1205 10:37:08.514436 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-service-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.514515 master-0 kubenswrapper[4752]: I1205 10:37:08.514488 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4825316a-ea9f-4d3d-838b-fa809a6e49c7-config\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.514545 master-0 kubenswrapper[4752]: I1205 10:37:08.514519 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e69ce9e-4e6f-4015-9ba6-5a7942570190-config\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:08.514581 master-0 kubenswrapper[4752]: I1205 10:37:08.514551 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c22d947f-a5b6-4f24-b142-dd201c46293b-serving-cert\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.514624 master-0 kubenswrapper[4752]: I1205 10:37:08.514577 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/283122ba-be1c-4516-bd0f-df41c13c098b-iptables-alerter-script\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:08.514624 master-0 kubenswrapper[4752]: I1205 10:37:08.514599 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fncz7\" (UniqueName: \"kubernetes.io/projected/926263c4-ec5b-41cb-9c30-0c88f636035f-kube-api-access-fncz7\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:08.514624 master-0 kubenswrapper[4752]: I1205 10:37:08.514620 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-profile-collector-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:08.514742 master-0 kubenswrapper[4752]: I1205 10:37:08.514644 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4825316a-ea9f-4d3d-838b-fa809a6e49c7-serving-cert\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.514742 master-0 kubenswrapper[4752]: I1205 10:37:08.514673 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv4zs\" (UniqueName: \"kubernetes.io/projected/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-kube-api-access-jv4zs\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:08.514742 master-0 kubenswrapper[4752]: I1205 10:37:08.514714 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-telemetry-config\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:08.514850 master-0 kubenswrapper[4752]: I1205 10:37:08.514747 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:08.514850 master-0 kubenswrapper[4752]: I1205 10:37:08.514769 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-config\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:08.514850 master-0 kubenswrapper[4752]: I1205 10:37:08.514792 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11f563d5-89bb-433c-956a-6d5d2492e8f1-serving-cert\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.514850 master-0 kubenswrapper[4752]: I1205 10:37:08.514840 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operand-assets\" (UniqueName: \"kubernetes.io/empty-dir/49051e6e-5a2f-45c8-bad0-374514a91c07-operand-assets\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:08.515006 master-0 kubenswrapper[4752]: I1205 10:37:08.514867 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjrgm\" (UniqueName: \"kubernetes.io/projected/87909f47-f2d7-46f8-a1c8-27336cdcce5d-kube-api-access-rjrgm\") pod \"csi-snapshot-controller-operator-6bc8656fdc-vd94f\" (UID: \"87909f47-f2d7-46f8-a1c8-27336cdcce5d\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:08.515006 master-0 kubenswrapper[4752]: I1205 10:37:08.514891 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11f563d5-89bb-433c-956a-6d5d2492e8f1-kube-api-access\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.515006 master-0 kubenswrapper[4752]: I1205 10:37:08.514915 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.515006 master-0 kubenswrapper[4752]: I1205 10:37:08.514938 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dtfn\" (UniqueName: \"kubernetes.io/projected/5a1bdc70-6412-47e0-8330-04d796cc8d55-kube-api-access-5dtfn\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:08.515006 master-0 kubenswrapper[4752]: I1205 10:37:08.514991 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7b29f89-e42d-4e53-ad14-05efdce933f0-trusted-ca\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.515188 master-0 kubenswrapper[4752]: I1205 10:37:08.515027 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:08.515231 master-0 kubenswrapper[4752]: I1205 10:37:08.515174 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c22d947f-a5b6-4f24-b142-dd201c46293b-available-featuregates\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.515411 master-0 kubenswrapper[4752]: I1205 10:37:08.515257 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98qt8\" (UniqueName: \"kubernetes.io/projected/c22d947f-a5b6-4f24-b142-dd201c46293b-kube-api-access-98qt8\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.515411 master-0 kubenswrapper[4752]: I1205 10:37:08.515353 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eb290494-a456-4f0e-9afc-f20abab1a1bf-trusted-ca\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:08.515411 master-0 kubenswrapper[4752]: I1205 10:37:08.515386 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/444f8808-e454-4015-9e20-429e715a08c7-config\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.515563 master-0 kubenswrapper[4752]: I1205 10:37:08.515444 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e69ce9e-4e6f-4015-9ba6-5a7942570190-serving-cert\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:08.515563 master-0 kubenswrapper[4752]: I1205 10:37:08.515474 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-config\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.515563 master-0 kubenswrapper[4752]: I1205 10:37:08.515497 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkmvj\" (UniqueName: \"kubernetes.io/projected/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-kube-api-access-fkmvj\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:08.515563 master-0 kubenswrapper[4752]: I1205 10:37:08.515523 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wfsv\" (UniqueName: \"kubernetes.io/projected/1e69ce9e-4e6f-4015-9ba6-5a7942570190-kube-api-access-7wfsv\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:08.515563 master-0 kubenswrapper[4752]: I1205 10:37:08.515547 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/283122ba-be1c-4516-bd0f-df41c13c098b-host-slash\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:08.515699 master-0 kubenswrapper[4752]: I1205 10:37:08.515567 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.515699 master-0 kubenswrapper[4752]: I1205 10:37:08.515589 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-client\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.515699 master-0 kubenswrapper[4752]: I1205 10:37:08.515613 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-config\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:08.515699 master-0 kubenswrapper[4752]: I1205 10:37:08.515636 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/444f8808-e454-4015-9e20-429e715a08c7-kube-api-access\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.515699 master-0 kubenswrapper[4752]: I1205 10:37:08.515658 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f563d5-89bb-433c-956a-6d5d2492e8f1-config\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.515699 master-0 kubenswrapper[4752]: I1205 10:37:08.515681 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4825316a-ea9f-4d3d-838b-fa809a6e49c7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.515861 master-0 kubenswrapper[4752]: I1205 10:37:08.515706 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vp9b\" (UniqueName: \"kubernetes.io/projected/49051e6e-5a2f-45c8-bad0-374514a91c07-kube-api-access-6vp9b\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:08.515861 master-0 kubenswrapper[4752]: I1205 10:37:08.515741 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twh85\" (UniqueName: \"kubernetes.io/projected/f7a08359-0379-4364-8b0c-ddb58ff605f4-kube-api-access-twh85\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.515861 master-0 kubenswrapper[4752]: I1205 10:37:08.515773 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-serving-cert\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:08.515861 master-0 kubenswrapper[4752]: I1205 10:37:08.515809 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-bound-sa-token\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.516092 master-0 kubenswrapper[4752]: I1205 10:37:08.515832 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-config\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.516137 master-0 kubenswrapper[4752]: I1205 10:37:08.516110 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-serving-cert\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.516166 master-0 kubenswrapper[4752]: I1205 10:37:08.516136 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:08.516166 master-0 kubenswrapper[4752]: I1205 10:37:08.516158 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f76d12f-5406-47e2-8337-2f50e35376d6-serving-cert\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:08.516219 master-0 kubenswrapper[4752]: I1205 10:37:08.516180 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-olm-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/49051e6e-5a2f-45c8-bad0-374514a91c07-cluster-olm-operator-serving-cert\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:08.516219 master-0 kubenswrapper[4752]: I1205 10:37:08.516202 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-serving-cert\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:08.516278 master-0 kubenswrapper[4752]: I1205 10:37:08.516229 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-trusted-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.516278 master-0 kubenswrapper[4752]: I1205 10:37:08.516251 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:08.516335 master-0 kubenswrapper[4752]: I1205 10:37:08.516276 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:08.516335 master-0 kubenswrapper[4752]: I1205 10:37:08.516299 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kb5c\" (UniqueName: \"kubernetes.io/projected/8d76404b-6d62-4a61-b6f6-0c8073eba198-kube-api-access-7kb5c\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:08.516335 master-0 kubenswrapper[4752]: I1205 10:37:08.516324 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-bound-sa-token\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.516410 master-0 kubenswrapper[4752]: I1205 10:37:08.516348 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f76d12f-5406-47e2-8337-2f50e35376d6-config\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:08.516410 master-0 kubenswrapper[4752]: I1205 10:37:08.516371 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/444f8808-e454-4015-9e20-429e715a08c7-serving-cert\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.516410 master-0 kubenswrapper[4752]: I1205 10:37:08.516392 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22676fac-b770-4937-9bee-7478bd1babb7-trusted-ca\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.617214 master-0 kubenswrapper[4752]: I1205 10:37:08.617139 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhjhk\" (UniqueName: \"kubernetes.io/projected/283122ba-be1c-4516-bd0f-df41c13c098b-kube-api-access-vhjhk\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:08.617214 master-0 kubenswrapper[4752]: I1205 10:37:08.617216 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lwgq\" (UniqueName: \"kubernetes.io/projected/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-kube-api-access-2lwgq\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:08.617459 master-0 kubenswrapper[4752]: I1205 10:37:08.617247 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:08.617459 master-0 kubenswrapper[4752]: I1205 10:37:08.617273 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn7rj\" (UniqueName: \"kubernetes.io/projected/eb290494-a456-4f0e-9afc-f20abab1a1bf-kube-api-access-qn7rj\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:08.617459 master-0 kubenswrapper[4752]: I1205 10:37:08.617296 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-serving-cert\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.617459 master-0 kubenswrapper[4752]: I1205 10:37:08.617317 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzwfq\" (UniqueName: \"kubernetes.io/projected/6f76d12f-5406-47e2-8337-2f50e35376d6-kube-api-access-fzwfq\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:08.617948 master-0 kubenswrapper[4752]: E1205 10:37:08.617904 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/performance-addon-operator-webhook-cert: secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:08.618138 master-0 kubenswrapper[4752]: I1205 10:37:08.618104 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkqz7\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-kube-api-access-vkqz7\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.618185 master-0 kubenswrapper[4752]: I1205 10:37:08.618148 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-service-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.618185 master-0 kubenswrapper[4752]: I1205 10:37:08.618173 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:08.618262 master-0 kubenswrapper[4752]: I1205 10:37:08.618199 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4825316a-ea9f-4d3d-838b-fa809a6e49c7-config\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.618262 master-0 kubenswrapper[4752]: I1205 10:37:08.618224 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2vd4\" (UniqueName: \"kubernetes.io/projected/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-kube-api-access-h2vd4\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:08.618262 master-0 kubenswrapper[4752]: I1205 10:37:08.618251 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-service-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.618362 master-0 kubenswrapper[4752]: I1205 10:37:08.618275 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e69ce9e-4e6f-4015-9ba6-5a7942570190-config\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:08.618362 master-0 kubenswrapper[4752]: I1205 10:37:08.618299 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c22d947f-a5b6-4f24-b142-dd201c46293b-serving-cert\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.618362 master-0 kubenswrapper[4752]: I1205 10:37:08.618325 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-profile-collector-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:08.618362 master-0 kubenswrapper[4752]: I1205 10:37:08.618346 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/283122ba-be1c-4516-bd0f-df41c13c098b-iptables-alerter-script\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:08.618587 master-0 kubenswrapper[4752]: I1205 10:37:08.618368 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fncz7\" (UniqueName: \"kubernetes.io/projected/926263c4-ec5b-41cb-9c30-0c88f636035f-kube-api-access-fncz7\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:08.618587 master-0 kubenswrapper[4752]: I1205 10:37:08.618392 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4825316a-ea9f-4d3d-838b-fa809a6e49c7-serving-cert\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.618587 master-0 kubenswrapper[4752]: I1205 10:37:08.618414 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv4zs\" (UniqueName: \"kubernetes.io/projected/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-kube-api-access-jv4zs\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:08.618587 master-0 kubenswrapper[4752]: I1205 10:37:08.618456 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-telemetry-config\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:08.618587 master-0 kubenswrapper[4752]: I1205 10:37:08.618481 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:08.618587 master-0 kubenswrapper[4752]: I1205 10:37:08.618502 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-config\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:08.618587 master-0 kubenswrapper[4752]: I1205 10:37:08.618537 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11f563d5-89bb-433c-956a-6d5d2492e8f1-serving-cert\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.618587 master-0 kubenswrapper[4752]: I1205 10:37:08.618577 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operand-assets\" (UniqueName: \"kubernetes.io/empty-dir/49051e6e-5a2f-45c8-bad0-374514a91c07-operand-assets\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:08.618793 master-0 kubenswrapper[4752]: I1205 10:37:08.618605 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.618793 master-0 kubenswrapper[4752]: I1205 10:37:08.618641 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjrgm\" (UniqueName: \"kubernetes.io/projected/87909f47-f2d7-46f8-a1c8-27336cdcce5d-kube-api-access-rjrgm\") pod \"csi-snapshot-controller-operator-6bc8656fdc-vd94f\" (UID: \"87909f47-f2d7-46f8-a1c8-27336cdcce5d\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:08.618793 master-0 kubenswrapper[4752]: I1205 10:37:08.618664 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11f563d5-89bb-433c-956a-6d5d2492e8f1-kube-api-access\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.618793 master-0 kubenswrapper[4752]: I1205 10:37:08.618691 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dtfn\" (UniqueName: \"kubernetes.io/projected/5a1bdc70-6412-47e0-8330-04d796cc8d55-kube-api-access-5dtfn\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:08.618793 master-0 kubenswrapper[4752]: I1205 10:37:08.618725 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7b29f89-e42d-4e53-ad14-05efdce933f0-trusted-ca\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.618793 master-0 kubenswrapper[4752]: I1205 10:37:08.618749 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:08.618793 master-0 kubenswrapper[4752]: I1205 10:37:08.618772 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c22d947f-a5b6-4f24-b142-dd201c46293b-available-featuregates\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.618974 master-0 kubenswrapper[4752]: I1205 10:37:08.618817 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98qt8\" (UniqueName: \"kubernetes.io/projected/c22d947f-a5b6-4f24-b142-dd201c46293b-kube-api-access-98qt8\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.618974 master-0 kubenswrapper[4752]: I1205 10:37:08.618846 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eb290494-a456-4f0e-9afc-f20abab1a1bf-trusted-ca\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:08.618974 master-0 kubenswrapper[4752]: I1205 10:37:08.618869 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/444f8808-e454-4015-9e20-429e715a08c7-config\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.618974 master-0 kubenswrapper[4752]: I1205 10:37:08.618896 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-config\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.618974 master-0 kubenswrapper[4752]: I1205 10:37:08.618921 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e69ce9e-4e6f-4015-9ba6-5a7942570190-serving-cert\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:08.618974 master-0 kubenswrapper[4752]: I1205 10:37:08.618945 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-service-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.619129 master-0 kubenswrapper[4752]: E1205 10:37:08.618951 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:09.118930471 +0000 UTC m=+110.659053880 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:08.619129 master-0 kubenswrapper[4752]: I1205 10:37:08.619005 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.619129 master-0 kubenswrapper[4752]: I1205 10:37:08.619040 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkmvj\" (UniqueName: \"kubernetes.io/projected/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-kube-api-access-fkmvj\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:08.619254 master-0 kubenswrapper[4752]: E1205 10:37:08.619223 4752 secret.go:189] Couldn't get secret openshift-image-registry/image-registry-operator-tls: secret "image-registry-operator-tls" not found Dec 05 10:37:08.619320 master-0 kubenswrapper[4752]: E1205 10:37:08.619304 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls podName:f7b29f89-e42d-4e53-ad14-05efdce933f0 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:09.119282211 +0000 UTC m=+110.659405570 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls") pod "cluster-image-registry-operator-6fb9f88b7-f29mb" (UID: "f7b29f89-e42d-4e53-ad14-05efdce933f0") : secret "image-registry-operator-tls" not found Dec 05 10:37:08.620944 master-0 kubenswrapper[4752]: E1205 10:37:08.619532 4752 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:08.620944 master-0 kubenswrapper[4752]: E1205 10:37:08.619588 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls podName:a722cda9-29a0-4b7f-8e1d-9a8950ed765a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:09.119574728 +0000 UTC m=+110.659698147 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-7ff994598c-kq8qr" (UID: "a722cda9-29a0-4b7f-8e1d-9a8950ed765a") : secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:08.620944 master-0 kubenswrapper[4752]: I1205 10:37:08.619752 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.620944 master-0 kubenswrapper[4752]: E1205 10:37:08.620192 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/node-tuning-operator-tls: secret "node-tuning-operator-tls" not found Dec 05 10:37:08.620944 master-0 kubenswrapper[4752]: E1205 10:37:08.620232 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:09.120220884 +0000 UTC m=+110.660344293 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-tuning-operator-tls" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "node-tuning-operator-tls" not found Dec 05 10:37:08.621377 master-0 kubenswrapper[4752]: I1205 10:37:08.621306 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operand-assets\" (UniqueName: \"kubernetes.io/empty-dir/49051e6e-5a2f-45c8-bad0-374514a91c07-operand-assets\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:08.622900 master-0 kubenswrapper[4752]: I1205 10:37:08.622029 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c22d947f-a5b6-4f24-b142-dd201c46293b-available-featuregates\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.622900 master-0 kubenswrapper[4752]: I1205 10:37:08.622830 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7b29f89-e42d-4e53-ad14-05efdce933f0-trusted-ca\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: E1205 10:37:08.623236 4752 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: secret "multus-admission-controller-secret" not found Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: E1205 10:37:08.623281 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs podName:8d76404b-6d62-4a61-b6f6-0c8073eba198 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:09.123269432 +0000 UTC m=+110.663392761 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs") pod "multus-admission-controller-7dfc5b745f-67rx7" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198") : secret "multus-admission-controller-secret" not found Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623412 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-telemetry-config\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623514 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wfsv\" (UniqueName: \"kubernetes.io/projected/1e69ce9e-4e6f-4015-9ba6-5a7942570190-kube-api-access-7wfsv\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623555 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/283122ba-be1c-4516-bd0f-df41c13c098b-host-slash\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623583 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-client\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623606 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-config\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623633 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/444f8808-e454-4015-9e20-429e715a08c7-kube-api-access\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623658 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vp9b\" (UniqueName: \"kubernetes.io/projected/49051e6e-5a2f-45c8-bad0-374514a91c07-kube-api-access-6vp9b\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623694 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f563d5-89bb-433c-956a-6d5d2492e8f1-config\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623724 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4825316a-ea9f-4d3d-838b-fa809a6e49c7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623752 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-serving-cert\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623781 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twh85\" (UniqueName: \"kubernetes.io/projected/f7a08359-0379-4364-8b0c-ddb58ff605f4-kube-api-access-twh85\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623791 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11f563d5-89bb-433c-956a-6d5d2492e8f1-serving-cert\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.625469 master-0 kubenswrapper[4752]: I1205 10:37:08.623810 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-serving-cert\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.623837 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-bound-sa-token\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.623867 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-config\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.623897 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.623927 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f76d12f-5406-47e2-8337-2f50e35376d6-serving-cert\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.623955 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-olm-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/49051e6e-5a2f-45c8-bad0-374514a91c07-cluster-olm-operator-serving-cert\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.623978 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-serving-cert\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.624007 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-trusted-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.624601 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-serving-cert\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.624647 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.624677 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.624704 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kb5c\" (UniqueName: \"kubernetes.io/projected/8d76404b-6d62-4a61-b6f6-0c8073eba198-kube-api-access-7kb5c\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.624730 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f76d12f-5406-47e2-8337-2f50e35376d6-config\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.624753 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-bound-sa-token\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.624771 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4825316a-ea9f-4d3d-838b-fa809a6e49c7-config\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.626112 master-0 kubenswrapper[4752]: I1205 10:37:08.624779 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/444f8808-e454-4015-9e20-429e715a08c7-serving-cert\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: I1205 10:37:08.624828 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22676fac-b770-4937-9bee-7478bd1babb7-trusted-ca\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: I1205 10:37:08.624862 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: I1205 10:37:08.624893 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/8c649a16-c187-412e-b5da-62a00bee38ab-kube-api-access-d8pv2\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: I1205 10:37:08.624919 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-trusted-ca\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: I1205 10:37:08.624941 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: I1205 10:37:08.624983 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fbss\" (UniqueName: \"kubernetes.io/projected/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-kube-api-access-2fbss\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: I1205 10:37:08.625009 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khfxz\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-kube-api-access-khfxz\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: I1205 10:37:08.625082 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-config\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: I1205 10:37:08.625468 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eb290494-a456-4f0e-9afc-f20abab1a1bf-trusted-ca\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: E1205 10:37:08.625529 4752 secret.go:189] Couldn't get secret openshift-dns-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: E1205 10:37:08.625590 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls podName:5a1bdc70-6412-47e0-8330-04d796cc8d55 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:09.125573601 +0000 UTC m=+110.665697000 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls") pod "dns-operator-7c56cf9b74-x6t9h" (UID: "5a1bdc70-6412-47e0-8330-04d796cc8d55") : secret "metrics-tls" not found Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: E1205 10:37:08.626036 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: secret "package-server-manager-serving-cert" not found Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: E1205 10:37:08.626068 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert podName:8c649a16-c187-412e-b5da-62a00bee38ab nodeName:}" failed. No retries permitted until 2025-12-05 10:37:09.126058093 +0000 UTC m=+110.666181522 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert") pod "package-server-manager-67477646d4-nm8cn" (UID: "8c649a16-c187-412e-b5da-62a00bee38ab") : secret "package-server-manager-serving-cert" not found Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: E1205 10:37:08.626122 4752 secret.go:189] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: secret "marketplace-operator-metrics" not found Dec 05 10:37:08.626661 master-0 kubenswrapper[4752]: E1205 10:37:08.626151 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics podName:926263c4-ec5b-41cb-9c30-0c88f636035f nodeName:}" failed. No retries permitted until 2025-12-05 10:37:09.126143076 +0000 UTC m=+110.666266485 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics") pod "marketplace-operator-f797b99b6-z9qcl" (UID: "926263c4-ec5b-41cb-9c30-0c88f636035f") : secret "marketplace-operator-metrics" not found Dec 05 10:37:08.627151 master-0 kubenswrapper[4752]: I1205 10:37:08.626939 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f76d12f-5406-47e2-8337-2f50e35376d6-config\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:08.627151 master-0 kubenswrapper[4752]: I1205 10:37:08.626929 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e69ce9e-4e6f-4015-9ba6-5a7942570190-config\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:08.627151 master-0 kubenswrapper[4752]: I1205 10:37:08.626960 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e69ce9e-4e6f-4015-9ba6-5a7942570190-serving-cert\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:08.627151 master-0 kubenswrapper[4752]: I1205 10:37:08.623800 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-profile-collector-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:08.627151 master-0 kubenswrapper[4752]: I1205 10:37:08.627016 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/283122ba-be1c-4516-bd0f-df41c13c098b-host-slash\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:08.627973 master-0 kubenswrapper[4752]: I1205 10:37:08.627878 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f563d5-89bb-433c-956a-6d5d2492e8f1-config\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.628150 master-0 kubenswrapper[4752]: I1205 10:37:08.628081 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-config\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:08.628610 master-0 kubenswrapper[4752]: I1205 10:37:08.628569 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/444f8808-e454-4015-9e20-429e715a08c7-serving-cert\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.628868 master-0 kubenswrapper[4752]: I1205 10:37:08.628836 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c22d947f-a5b6-4f24-b142-dd201c46293b-serving-cert\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.629523 master-0 kubenswrapper[4752]: I1205 10:37:08.629488 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/444f8808-e454-4015-9e20-429e715a08c7-config\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.630971 master-0 kubenswrapper[4752]: I1205 10:37:08.630926 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-service-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.631853 master-0 kubenswrapper[4752]: E1205 10:37:08.631138 4752 secret.go:189] Couldn't get secret openshift-ingress-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:08.631853 master-0 kubenswrapper[4752]: I1205 10:37:08.631216 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-serving-cert\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.631853 master-0 kubenswrapper[4752]: E1205 10:37:08.631309 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: secret "olm-operator-serving-cert" not found Dec 05 10:37:08.631853 master-0 kubenswrapper[4752]: E1205 10:37:08.631359 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert podName:46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:09.131340048 +0000 UTC m=+110.671463577 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert") pod "olm-operator-7cd7dbb44c-d25sk" (UID: "46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082") : secret "olm-operator-serving-cert" not found Dec 05 10:37:08.631853 master-0 kubenswrapper[4752]: I1205 10:37:08.631449 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-client\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.631853 master-0 kubenswrapper[4752]: E1205 10:37:08.631494 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls podName:22676fac-b770-4937-9bee-7478bd1babb7 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:09.131482952 +0000 UTC m=+110.671606391 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls") pod "ingress-operator-8649c48786-cgt5x" (UID: "22676fac-b770-4937-9bee-7478bd1babb7") : secret "metrics-tls" not found Dec 05 10:37:08.632763 master-0 kubenswrapper[4752]: I1205 10:37:08.632467 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22676fac-b770-4937-9bee-7478bd1babb7-trusted-ca\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.632763 master-0 kubenswrapper[4752]: I1205 10:37:08.632691 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-trusted-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.632885 master-0 kubenswrapper[4752]: I1205 10:37:08.632778 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-config\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.633005 master-0 kubenswrapper[4752]: I1205 10:37:08.632962 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-trusted-ca\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:08.633298 master-0 kubenswrapper[4752]: I1205 10:37:08.633256 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/283122ba-be1c-4516-bd0f-df41c13c098b-iptables-alerter-script\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:08.634345 master-0 kubenswrapper[4752]: I1205 10:37:08.634307 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4825316a-ea9f-4d3d-838b-fa809a6e49c7-serving-cert\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.636897 master-0 kubenswrapper[4752]: I1205 10:37:08.636540 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-olm-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/49051e6e-5a2f-45c8-bad0-374514a91c07-cluster-olm-operator-serving-cert\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:08.638521 master-0 kubenswrapper[4752]: I1205 10:37:08.638390 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-serving-cert\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:08.641713 master-0 kubenswrapper[4752]: I1205 10:37:08.641662 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn7rj\" (UniqueName: \"kubernetes.io/projected/eb290494-a456-4f0e-9afc-f20abab1a1bf-kube-api-access-qn7rj\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:08.641973 master-0 kubenswrapper[4752]: I1205 10:37:08.641933 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98qt8\" (UniqueName: \"kubernetes.io/projected/c22d947f-a5b6-4f24-b142-dd201c46293b-kube-api-access-98qt8\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.642121 master-0 kubenswrapper[4752]: I1205 10:37:08.642082 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkmvj\" (UniqueName: \"kubernetes.io/projected/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-kube-api-access-fkmvj\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:08.642913 master-0 kubenswrapper[4752]: I1205 10:37:08.642873 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f76d12f-5406-47e2-8337-2f50e35376d6-serving-cert\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:08.643444 master-0 kubenswrapper[4752]: I1205 10:37:08.643399 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhjhk\" (UniqueName: \"kubernetes.io/projected/283122ba-be1c-4516-bd0f-df41c13c098b-kube-api-access-vhjhk\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:08.643741 master-0 kubenswrapper[4752]: I1205 10:37:08.643703 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzwfq\" (UniqueName: \"kubernetes.io/projected/6f76d12f-5406-47e2-8337-2f50e35376d6-kube-api-access-fzwfq\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:08.644114 master-0 kubenswrapper[4752]: I1205 10:37:08.644062 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-config\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:08.652937 master-0 kubenswrapper[4752]: I1205 10:37:08.649810 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-serving-cert\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:08.652937 master-0 kubenswrapper[4752]: I1205 10:37:08.650725 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lwgq\" (UniqueName: \"kubernetes.io/projected/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-kube-api-access-2lwgq\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:08.654307 master-0 kubenswrapper[4752]: I1205 10:37:08.653286 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11f563d5-89bb-433c-956a-6d5d2492e8f1-kube-api-access\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.655015 master-0 kubenswrapper[4752]: I1205 10:37:08.654957 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wfsv\" (UniqueName: \"kubernetes.io/projected/1e69ce9e-4e6f-4015-9ba6-5a7942570190-kube-api-access-7wfsv\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:08.656848 master-0 kubenswrapper[4752]: I1205 10:37:08.656793 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv4zs\" (UniqueName: \"kubernetes.io/projected/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-kube-api-access-jv4zs\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:08.656848 master-0 kubenswrapper[4752]: I1205 10:37:08.656836 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/444f8808-e454-4015-9e20-429e715a08c7-kube-api-access\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.658020 master-0 kubenswrapper[4752]: I1205 10:37:08.657977 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkqz7\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-kube-api-access-vkqz7\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.658677 master-0 kubenswrapper[4752]: I1205 10:37:08.658642 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vp9b\" (UniqueName: \"kubernetes.io/projected/49051e6e-5a2f-45c8-bad0-374514a91c07-kube-api-access-6vp9b\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:08.674912 master-0 kubenswrapper[4752]: I1205 10:37:08.674879 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4825316a-ea9f-4d3d-838b-fa809a6e49c7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.676061 master-0 kubenswrapper[4752]: I1205 10:37:08.675576 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.693404 master-0 kubenswrapper[4752]: I1205 10:37:08.693346 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kb5c\" (UniqueName: \"kubernetes.io/projected/8d76404b-6d62-4a61-b6f6-0c8073eba198-kube-api-access-7kb5c\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:08.708736 master-0 kubenswrapper[4752]: I1205 10:37:08.708690 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.712452 master-0 kubenswrapper[4752]: I1205 10:37:08.712270 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khfxz\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-kube-api-access-khfxz\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.730862 master-0 kubenswrapper[4752]: I1205 10:37:08.730396 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.733375 master-0 kubenswrapper[4752]: I1205 10:37:08.733337 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twh85\" (UniqueName: \"kubernetes.io/projected/f7a08359-0379-4364-8b0c-ddb58ff605f4-kube-api-access-twh85\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.765913 master-0 kubenswrapper[4752]: I1205 10:37:08.755520 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-bound-sa-token\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:08.778936 master-0 kubenswrapper[4752]: I1205 10:37:08.778510 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/8c649a16-c187-412e-b5da-62a00bee38ab-kube-api-access-d8pv2\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:08.787231 master-0 kubenswrapper[4752]: I1205 10:37:08.787184 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.802590 master-0 kubenswrapper[4752]: I1205 10:37:08.802539 4752 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2a62edc-8f1c-43be-b355-8484bbebac86" path="/var/lib/kubelet/pods/b2a62edc-8f1c-43be-b355-8484bbebac86/volumes" Dec 05 10:37:08.802877 master-0 kubenswrapper[4752]: I1205 10:37:08.802839 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-bound-sa-token\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:08.817261 master-0 kubenswrapper[4752]: I1205 10:37:08.817146 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fbss\" (UniqueName: \"kubernetes.io/projected/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-kube-api-access-2fbss\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.821826 master-0 kubenswrapper[4752]: I1205 10:37:08.821534 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:08.832992 master-0 kubenswrapper[4752]: W1205 10:37:08.832937 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod283122ba_be1c_4516_bd0f_df41c13c098b.slice/crio-b44ac27d1b7f531e11c3f4d352d89c4a7679a795ec64417e43c87c5d1e725e90 WatchSource:0}: Error finding container b44ac27d1b7f531e11c3f4d352d89c4a7679a795ec64417e43c87c5d1e725e90: Status 404 returned error can't find the container with id b44ac27d1b7f531e11c3f4d352d89c4a7679a795ec64417e43c87c5d1e725e90 Dec 05 10:37:08.837788 master-0 kubenswrapper[4752]: I1205 10:37:08.837732 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fncz7\" (UniqueName: \"kubernetes.io/projected/926263c4-ec5b-41cb-9c30-0c88f636035f-kube-api-access-fncz7\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:08.844201 master-0 kubenswrapper[4752]: I1205 10:37:08.844158 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: E1205 10:37:08.847325 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator_444f8808-e454-4015-9e20-429e715a08c7_0(2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330): error adding pod openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330" Netns:"/var/run/netns/f9ad0b0b-d35d-4d49-9b97-5f1fe1949d54" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-apiserver-operator;K8S_POD_NAME=kube-apiserver-operator-765d9ff747-p57fl;K8S_POD_INFRA_CONTAINER_ID=2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330;K8S_POD_UID=444f8808-e454-4015-9e20-429e715a08c7" Path:"" ERRORED: error configuring pod [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl] networking: [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl/444f8808-e454-4015-9e20-429e715a08c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: > Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: E1205 10:37:08.847388 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator_444f8808-e454-4015-9e20-429e715a08c7_0(2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330): error adding pod openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330" Netns:"/var/run/netns/f9ad0b0b-d35d-4d49-9b97-5f1fe1949d54" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-apiserver-operator;K8S_POD_NAME=kube-apiserver-operator-765d9ff747-p57fl;K8S_POD_INFRA_CONTAINER_ID=2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330;K8S_POD_UID=444f8808-e454-4015-9e20-429e715a08c7" Path:"" ERRORED: error configuring pod [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl] networking: [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl/444f8808-e454-4015-9e20-429e715a08c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: > pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: E1205 10:37:08.847409 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator_444f8808-e454-4015-9e20-429e715a08c7_0(2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330): error adding pod openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330" Netns:"/var/run/netns/f9ad0b0b-d35d-4d49-9b97-5f1fe1949d54" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-apiserver-operator;K8S_POD_NAME=kube-apiserver-operator-765d9ff747-p57fl;K8S_POD_INFRA_CONTAINER_ID=2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330;K8S_POD_UID=444f8808-e454-4015-9e20-429e715a08c7" Path:"" ERRORED: error configuring pod [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl] networking: [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl/444f8808-e454-4015-9e20-429e715a08c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.847501 master-0 kubenswrapper[4752]: > pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:08.847903 master-0 kubenswrapper[4752]: E1205 10:37:08.847487 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator(444f8808-e454-4015-9e20-429e715a08c7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator(444f8808-e454-4015-9e20-429e715a08c7)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator_444f8808-e454-4015-9e20-429e715a08c7_0(2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330): error adding pod openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330\\\" Netns:\\\"/var/run/netns/f9ad0b0b-d35d-4d49-9b97-5f1fe1949d54\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-apiserver-operator;K8S_POD_NAME=kube-apiserver-operator-765d9ff747-p57fl;K8S_POD_INFRA_CONTAINER_ID=2788d56fa426b4c865f7d79c8d59769775aa83f7f48d9d978cf6a2b76ece0330;K8S_POD_UID=444f8808-e454-4015-9e20-429e715a08c7\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl] networking: [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl/444f8808-e454-4015-9e20-429e715a08c7:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" podUID="444f8808-e454-4015-9e20-429e715a08c7" Dec 05 10:37:08.853397 master-0 kubenswrapper[4752]: I1205 10:37:08.853047 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:08.860030 master-0 kubenswrapper[4752]: I1205 10:37:08.859985 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dtfn\" (UniqueName: \"kubernetes.io/projected/5a1bdc70-6412-47e0-8330-04d796cc8d55-kube-api-access-5dtfn\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:08.867637 master-0 kubenswrapper[4752]: I1205 10:37:08.866751 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:08.875172 master-0 kubenswrapper[4752]: I1205 10:37:08.875109 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjrgm\" (UniqueName: \"kubernetes.io/projected/87909f47-f2d7-46f8-a1c8-27336cdcce5d-kube-api-access-rjrgm\") pod \"csi-snapshot-controller-operator-6bc8656fdc-vd94f\" (UID: \"87909f47-f2d7-46f8-a1c8-27336cdcce5d\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: E1205 10:37:08.880924 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator_11f563d5-89bb-433c-956a-6d5d2492e8f1_0(e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff): error adding pod openshift-kube-controller-manager-operator_kube-controller-manager-operator-848f645654-rmdb8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff" Netns:"/var/run/netns/5e13bc21-7f5b-401d-ac04-f030a1811db8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager-operator;K8S_POD_NAME=kube-controller-manager-operator-848f645654-rmdb8;K8S_POD_INFRA_CONTAINER_ID=e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff;K8S_POD_UID=11f563d5-89bb-433c-956a-6d5d2492e8f1" Path:"" ERRORED: error configuring pod [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8] networking: [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8/11f563d5-89bb-433c-956a-6d5d2492e8f1:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: > Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: I1205 10:37:08.880961 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: E1205 10:37:08.880985 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator_11f563d5-89bb-433c-956a-6d5d2492e8f1_0(e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff): error adding pod openshift-kube-controller-manager-operator_kube-controller-manager-operator-848f645654-rmdb8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff" Netns:"/var/run/netns/5e13bc21-7f5b-401d-ac04-f030a1811db8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager-operator;K8S_POD_NAME=kube-controller-manager-operator-848f645654-rmdb8;K8S_POD_INFRA_CONTAINER_ID=e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff;K8S_POD_UID=11f563d5-89bb-433c-956a-6d5d2492e8f1" Path:"" ERRORED: error configuring pod [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8] networking: [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8/11f563d5-89bb-433c-956a-6d5d2492e8f1:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: > pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: E1205 10:37:08.881004 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator_11f563d5-89bb-433c-956a-6d5d2492e8f1_0(e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff): error adding pod openshift-kube-controller-manager-operator_kube-controller-manager-operator-848f645654-rmdb8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff" Netns:"/var/run/netns/5e13bc21-7f5b-401d-ac04-f030a1811db8" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager-operator;K8S_POD_NAME=kube-controller-manager-operator-848f645654-rmdb8;K8S_POD_INFRA_CONTAINER_ID=e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff;K8S_POD_UID=11f563d5-89bb-433c-956a-6d5d2492e8f1" Path:"" ERRORED: error configuring pod [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8] networking: [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8/11f563d5-89bb-433c-956a-6d5d2492e8f1:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.881319 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.881904 master-0 kubenswrapper[4752]: > pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:08.881904 master-0 kubenswrapper[4752]: E1205 10:37:08.881049 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator(11f563d5-89bb-433c-956a-6d5d2492e8f1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator(11f563d5-89bb-433c-956a-6d5d2492e8f1)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator_11f563d5-89bb-433c-956a-6d5d2492e8f1_0(e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff): error adding pod openshift-kube-controller-manager-operator_kube-controller-manager-operator-848f645654-rmdb8 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff\\\" Netns:\\\"/var/run/netns/5e13bc21-7f5b-401d-ac04-f030a1811db8\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager-operator;K8S_POD_NAME=kube-controller-manager-operator-848f645654-rmdb8;K8S_POD_INFRA_CONTAINER_ID=e0abdf47379f1e4f4273c269024eca936557d68b103bfe71c1226905eb88e5ff;K8S_POD_UID=11f563d5-89bb-433c-956a-6d5d2492e8f1\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8] networking: [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8/11f563d5-89bb-433c-956a-6d5d2492e8f1:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" podUID="11f563d5-89bb-433c-956a-6d5d2492e8f1" Dec 05 10:37:08.893585 master-0 kubenswrapper[4752]: I1205 10:37:08.893549 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2vd4\" (UniqueName: \"kubernetes.io/projected/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-kube-api-access-h2vd4\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:08.897441 master-0 kubenswrapper[4752]: I1205 10:37:08.897391 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:08.928450 master-0 kubenswrapper[4752]: I1205 10:37:08.926712 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:08.959784 master-0 kubenswrapper[4752]: E1205 10:37:08.959707 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:08.959784 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator_4825316a-ea9f-4d3d-838b-fa809a6e49c7_0(1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f): error adding pod openshift-kube-scheduler-operator_openshift-kube-scheduler-operator-5f85974995-dwh5t to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f" Netns:"/var/run/netns/70c7ba8b-77d1-4864-84e1-9192e7f66f5f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-scheduler-operator;K8S_POD_NAME=openshift-kube-scheduler-operator-5f85974995-dwh5t;K8S_POD_INFRA_CONTAINER_ID=1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f;K8S_POD_UID=4825316a-ea9f-4d3d-838b-fa809a6e49c7" Path:"" ERRORED: error configuring pod [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t] networking: [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t/4825316a-ea9f-4d3d-838b-fa809a6e49c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.959784 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.959784 master-0 kubenswrapper[4752]: > Dec 05 10:37:08.960025 master-0 kubenswrapper[4752]: E1205 10:37:08.959792 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:08.960025 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator_4825316a-ea9f-4d3d-838b-fa809a6e49c7_0(1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f): error adding pod openshift-kube-scheduler-operator_openshift-kube-scheduler-operator-5f85974995-dwh5t to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f" Netns:"/var/run/netns/70c7ba8b-77d1-4864-84e1-9192e7f66f5f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-scheduler-operator;K8S_POD_NAME=openshift-kube-scheduler-operator-5f85974995-dwh5t;K8S_POD_INFRA_CONTAINER_ID=1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f;K8S_POD_UID=4825316a-ea9f-4d3d-838b-fa809a6e49c7" Path:"" ERRORED: error configuring pod [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t] networking: [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t/4825316a-ea9f-4d3d-838b-fa809a6e49c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.960025 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.960025 master-0 kubenswrapper[4752]: > pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.960025 master-0 kubenswrapper[4752]: E1205 10:37:08.959827 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:08.960025 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator_4825316a-ea9f-4d3d-838b-fa809a6e49c7_0(1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f): error adding pod openshift-kube-scheduler-operator_openshift-kube-scheduler-operator-5f85974995-dwh5t to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f" Netns:"/var/run/netns/70c7ba8b-77d1-4864-84e1-9192e7f66f5f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-scheduler-operator;K8S_POD_NAME=openshift-kube-scheduler-operator-5f85974995-dwh5t;K8S_POD_INFRA_CONTAINER_ID=1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f;K8S_POD_UID=4825316a-ea9f-4d3d-838b-fa809a6e49c7" Path:"" ERRORED: error configuring pod [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t] networking: [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t/4825316a-ea9f-4d3d-838b-fa809a6e49c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.960025 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.960025 master-0 kubenswrapper[4752]: > pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:08.960218 master-0 kubenswrapper[4752]: E1205 10:37:08.959940 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator(4825316a-ea9f-4d3d-838b-fa809a6e49c7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator(4825316a-ea9f-4d3d-838b-fa809a6e49c7)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator_4825316a-ea9f-4d3d-838b-fa809a6e49c7_0(1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f): error adding pod openshift-kube-scheduler-operator_openshift-kube-scheduler-operator-5f85974995-dwh5t to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f\\\" Netns:\\\"/var/run/netns/70c7ba8b-77d1-4864-84e1-9192e7f66f5f\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-scheduler-operator;K8S_POD_NAME=openshift-kube-scheduler-operator-5f85974995-dwh5t;K8S_POD_INFRA_CONTAINER_ID=1478dbbbd3cfbc5cb76d924765ade9dd3e1187eeedc74f74e2dce5250f4a065f;K8S_POD_UID=4825316a-ea9f-4d3d-838b-fa809a6e49c7\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t] networking: [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t/4825316a-ea9f-4d3d-838b-fa809a6e49c7:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" podUID="4825316a-ea9f-4d3d-838b-fa809a6e49c7" Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: E1205 10:37:08.979779 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator_c22d947f-a5b6-4f24-b142-dd201c46293b_0(8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f): error adding pod openshift-config-operator_openshift-config-operator-68758cbcdb-dnpcv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f" Netns:"/var/run/netns/f360a8d6-c9c4-4fa5-a397-5d76c6cd3aba" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-config-operator;K8S_POD_NAME=openshift-config-operator-68758cbcdb-dnpcv;K8S_POD_INFRA_CONTAINER_ID=8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f;K8S_POD_UID=c22d947f-a5b6-4f24-b142-dd201c46293b" Path:"" ERRORED: error configuring pod [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv] networking: [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv/c22d947f-a5b6-4f24-b142-dd201c46293b:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: > Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: E1205 10:37:08.979835 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator_c22d947f-a5b6-4f24-b142-dd201c46293b_0(8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f): error adding pod openshift-config-operator_openshift-config-operator-68758cbcdb-dnpcv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f" Netns:"/var/run/netns/f360a8d6-c9c4-4fa5-a397-5d76c6cd3aba" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-config-operator;K8S_POD_NAME=openshift-config-operator-68758cbcdb-dnpcv;K8S_POD_INFRA_CONTAINER_ID=8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f;K8S_POD_UID=c22d947f-a5b6-4f24-b142-dd201c46293b" Path:"" ERRORED: error configuring pod [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv] networking: [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv/c22d947f-a5b6-4f24-b142-dd201c46293b:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: > pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: E1205 10:37:08.979858 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator_c22d947f-a5b6-4f24-b142-dd201c46293b_0(8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f): error adding pod openshift-config-operator_openshift-config-operator-68758cbcdb-dnpcv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f" Netns:"/var/run/netns/f360a8d6-c9c4-4fa5-a397-5d76c6cd3aba" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-config-operator;K8S_POD_NAME=openshift-config-operator-68758cbcdb-dnpcv;K8S_POD_INFRA_CONTAINER_ID=8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f;K8S_POD_UID=c22d947f-a5b6-4f24-b142-dd201c46293b" Path:"" ERRORED: error configuring pod [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv] networking: [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv/c22d947f-a5b6-4f24-b142-dd201c46293b:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:08.980225 master-0 kubenswrapper[4752]: > pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:08.980713 master-0 kubenswrapper[4752]: E1205 10:37:08.979930 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator(c22d947f-a5b6-4f24-b142-dd201c46293b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator(c22d947f-a5b6-4f24-b142-dd201c46293b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator_c22d947f-a5b6-4f24-b142-dd201c46293b_0(8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f): error adding pod openshift-config-operator_openshift-config-operator-68758cbcdb-dnpcv to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f\\\" Netns:\\\"/var/run/netns/f360a8d6-c9c4-4fa5-a397-5d76c6cd3aba\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-config-operator;K8S_POD_NAME=openshift-config-operator-68758cbcdb-dnpcv;K8S_POD_INFRA_CONTAINER_ID=8a52876c8b6af56e50a91edc7e76be09107ac8a910e3419ac5a6a5819d09f72f;K8S_POD_UID=c22d947f-a5b6-4f24-b142-dd201c46293b\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv] networking: [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv/c22d947f-a5b6-4f24-b142-dd201c46293b:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" podUID="c22d947f-a5b6-4f24-b142-dd201c46293b" Dec 05 10:37:09.025377 master-0 kubenswrapper[4752]: E1205 10:37:09.025315 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:09.025377 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator_49051e6e-5a2f-45c8-bad0-374514a91c07_0(53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58): error adding pod openshift-cluster-olm-operator_cluster-olm-operator-56fcb6cc5f-m6p27 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58" Netns:"/var/run/netns/6fb84fe3-2eaf-4c60-88d7-5e7015f3a080" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-olm-operator;K8S_POD_NAME=cluster-olm-operator-56fcb6cc5f-m6p27;K8S_POD_INFRA_CONTAINER_ID=53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58;K8S_POD_UID=49051e6e-5a2f-45c8-bad0-374514a91c07" Path:"" ERRORED: error configuring pod [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27] networking: [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27/49051e6e-5a2f-45c8-bad0-374514a91c07:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.025377 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.025377 master-0 kubenswrapper[4752]: > Dec 05 10:37:09.025543 master-0 kubenswrapper[4752]: E1205 10:37:09.025387 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:09.025543 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator_49051e6e-5a2f-45c8-bad0-374514a91c07_0(53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58): error adding pod openshift-cluster-olm-operator_cluster-olm-operator-56fcb6cc5f-m6p27 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58" Netns:"/var/run/netns/6fb84fe3-2eaf-4c60-88d7-5e7015f3a080" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-olm-operator;K8S_POD_NAME=cluster-olm-operator-56fcb6cc5f-m6p27;K8S_POD_INFRA_CONTAINER_ID=53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58;K8S_POD_UID=49051e6e-5a2f-45c8-bad0-374514a91c07" Path:"" ERRORED: error configuring pod [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27] networking: [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27/49051e6e-5a2f-45c8-bad0-374514a91c07:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.025543 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.025543 master-0 kubenswrapper[4752]: > pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:09.025543 master-0 kubenswrapper[4752]: E1205 10:37:09.025433 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:09.025543 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator_49051e6e-5a2f-45c8-bad0-374514a91c07_0(53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58): error adding pod openshift-cluster-olm-operator_cluster-olm-operator-56fcb6cc5f-m6p27 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58" Netns:"/var/run/netns/6fb84fe3-2eaf-4c60-88d7-5e7015f3a080" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-olm-operator;K8S_POD_NAME=cluster-olm-operator-56fcb6cc5f-m6p27;K8S_POD_INFRA_CONTAINER_ID=53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58;K8S_POD_UID=49051e6e-5a2f-45c8-bad0-374514a91c07" Path:"" ERRORED: error configuring pod [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27] networking: [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27/49051e6e-5a2f-45c8-bad0-374514a91c07:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.025543 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.025543 master-0 kubenswrapper[4752]: > pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:09.025737 master-0 kubenswrapper[4752]: E1205 10:37:09.025506 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator(49051e6e-5a2f-45c8-bad0-374514a91c07)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator(49051e6e-5a2f-45c8-bad0-374514a91c07)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator_49051e6e-5a2f-45c8-bad0-374514a91c07_0(53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58): error adding pod openshift-cluster-olm-operator_cluster-olm-operator-56fcb6cc5f-m6p27 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58\\\" Netns:\\\"/var/run/netns/6fb84fe3-2eaf-4c60-88d7-5e7015f3a080\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-olm-operator;K8S_POD_NAME=cluster-olm-operator-56fcb6cc5f-m6p27;K8S_POD_INFRA_CONTAINER_ID=53d29421c1702504b0246daab771474e4e996bdd186b2c3633464b1adf78ab58;K8S_POD_UID=49051e6e-5a2f-45c8-bad0-374514a91c07\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27] networking: [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27/49051e6e-5a2f-45c8-bad0-374514a91c07:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" podUID="49051e6e-5a2f-45c8-bad0-374514a91c07" Dec 05 10:37:09.032050 master-0 kubenswrapper[4752]: E1205 10:37:09.031996 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:09.032050 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator_d95a56ba-c940-4e3e-aed6-d8c04f1871b6_0(aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d): error adding pod openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d" Netns:"/var/run/netns/a6afdfd5-614a-4f71-8a84-5b57041533ef" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication-operator;K8S_POD_NAME=authentication-operator-6c968fdfdf-t7sl8;K8S_POD_INFRA_CONTAINER_ID=aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d;K8S_POD_UID=d95a56ba-c940-4e3e-aed6-d8c04f1871b6" Path:"" ERRORED: error configuring pod [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8] networking: [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8/d95a56ba-c940-4e3e-aed6-d8c04f1871b6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.032050 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.032050 master-0 kubenswrapper[4752]: > Dec 05 10:37:09.032234 master-0 kubenswrapper[4752]: E1205 10:37:09.032073 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:09.032234 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator_d95a56ba-c940-4e3e-aed6-d8c04f1871b6_0(aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d): error adding pod openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d" Netns:"/var/run/netns/a6afdfd5-614a-4f71-8a84-5b57041533ef" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication-operator;K8S_POD_NAME=authentication-operator-6c968fdfdf-t7sl8;K8S_POD_INFRA_CONTAINER_ID=aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d;K8S_POD_UID=d95a56ba-c940-4e3e-aed6-d8c04f1871b6" Path:"" ERRORED: error configuring pod [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8] networking: [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8/d95a56ba-c940-4e3e-aed6-d8c04f1871b6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.032234 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.032234 master-0 kubenswrapper[4752]: > pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:09.032234 master-0 kubenswrapper[4752]: E1205 10:37:09.032092 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:09.032234 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator_d95a56ba-c940-4e3e-aed6-d8c04f1871b6_0(aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d): error adding pod openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d" Netns:"/var/run/netns/a6afdfd5-614a-4f71-8a84-5b57041533ef" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication-operator;K8S_POD_NAME=authentication-operator-6c968fdfdf-t7sl8;K8S_POD_INFRA_CONTAINER_ID=aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d;K8S_POD_UID=d95a56ba-c940-4e3e-aed6-d8c04f1871b6" Path:"" ERRORED: error configuring pod [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8] networking: [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8/d95a56ba-c940-4e3e-aed6-d8c04f1871b6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.032234 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.032234 master-0 kubenswrapper[4752]: > pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:09.032832 master-0 kubenswrapper[4752]: E1205 10:37:09.032153 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator(d95a56ba-c940-4e3e-aed6-d8c04f1871b6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator(d95a56ba-c940-4e3e-aed6-d8c04f1871b6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator_d95a56ba-c940-4e3e-aed6-d8c04f1871b6_0(aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d): error adding pod openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d\\\" Netns:\\\"/var/run/netns/a6afdfd5-614a-4f71-8a84-5b57041533ef\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication-operator;K8S_POD_NAME=authentication-operator-6c968fdfdf-t7sl8;K8S_POD_INFRA_CONTAINER_ID=aa0846858f773d130ce5a631569e5ed898f89b39a29a2f10a3d9985541255d1d;K8S_POD_UID=d95a56ba-c940-4e3e-aed6-d8c04f1871b6\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8] networking: [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8/d95a56ba-c940-4e3e-aed6-d8c04f1871b6:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" podUID="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" Dec 05 10:37:09.043403 master-0 kubenswrapper[4752]: E1205 10:37:09.043346 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:09.043403 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator_f7a08359-0379-4364-8b0c-ddb58ff605f4_0(750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0): error adding pod openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0" Netns:"/var/run/netns/3a3b18b9-f2b6-4219-91d8-66919b044171" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-etcd-operator;K8S_POD_NAME=etcd-operator-5bf4d88c6f-n8t5c;K8S_POD_INFRA_CONTAINER_ID=750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0;K8S_POD_UID=f7a08359-0379-4364-8b0c-ddb58ff605f4" Path:"" ERRORED: error configuring pod [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c] networking: [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c/f7a08359-0379-4364-8b0c-ddb58ff605f4:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.043403 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.043403 master-0 kubenswrapper[4752]: > Dec 05 10:37:09.043624 master-0 kubenswrapper[4752]: E1205 10:37:09.043434 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:09.043624 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator_f7a08359-0379-4364-8b0c-ddb58ff605f4_0(750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0): error adding pod openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0" Netns:"/var/run/netns/3a3b18b9-f2b6-4219-91d8-66919b044171" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-etcd-operator;K8S_POD_NAME=etcd-operator-5bf4d88c6f-n8t5c;K8S_POD_INFRA_CONTAINER_ID=750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0;K8S_POD_UID=f7a08359-0379-4364-8b0c-ddb58ff605f4" Path:"" ERRORED: error configuring pod [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c] networking: [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c/f7a08359-0379-4364-8b0c-ddb58ff605f4:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.043624 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.043624 master-0 kubenswrapper[4752]: > pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:09.043624 master-0 kubenswrapper[4752]: E1205 10:37:09.043460 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:09.043624 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator_f7a08359-0379-4364-8b0c-ddb58ff605f4_0(750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0): error adding pod openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0" Netns:"/var/run/netns/3a3b18b9-f2b6-4219-91d8-66919b044171" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-etcd-operator;K8S_POD_NAME=etcd-operator-5bf4d88c6f-n8t5c;K8S_POD_INFRA_CONTAINER_ID=750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0;K8S_POD_UID=f7a08359-0379-4364-8b0c-ddb58ff605f4" Path:"" ERRORED: error configuring pod [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c] networking: [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c/f7a08359-0379-4364-8b0c-ddb58ff605f4:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.043624 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.043624 master-0 kubenswrapper[4752]: > pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:09.043624 master-0 kubenswrapper[4752]: E1205 10:37:09.043522 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator(f7a08359-0379-4364-8b0c-ddb58ff605f4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator(f7a08359-0379-4364-8b0c-ddb58ff605f4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator_f7a08359-0379-4364-8b0c-ddb58ff605f4_0(750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0): error adding pod openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0\\\" Netns:\\\"/var/run/netns/3a3b18b9-f2b6-4219-91d8-66919b044171\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-etcd-operator;K8S_POD_NAME=etcd-operator-5bf4d88c6f-n8t5c;K8S_POD_INFRA_CONTAINER_ID=750661c57588737ef708cd550d7ce5c96c84e2bbd40104b467e0d65b8655aec0;K8S_POD_UID=f7a08359-0379-4364-8b0c-ddb58ff605f4\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c] networking: [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c/f7a08359-0379-4364-8b0c-ddb58ff605f4:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" podUID="f7a08359-0379-4364-8b0c-ddb58ff605f4" Dec 05 10:37:09.068187 master-0 kubenswrapper[4752]: E1205 10:37:09.068141 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:09.068187 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator_6f76d12f-5406-47e2-8337-2f50e35376d6_0(7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe): error adding pod openshift-apiserver-operator_openshift-apiserver-operator-7bf7f6b755-hdjv7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe" Netns:"/var/run/netns/895de962-b7a8-4b19-b798-ff7111d62b48" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-apiserver-operator;K8S_POD_NAME=openshift-apiserver-operator-7bf7f6b755-hdjv7;K8S_POD_INFRA_CONTAINER_ID=7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe;K8S_POD_UID=6f76d12f-5406-47e2-8337-2f50e35376d6" Path:"" ERRORED: error configuring pod [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7] networking: [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7/6f76d12f-5406-47e2-8337-2f50e35376d6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.068187 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.068187 master-0 kubenswrapper[4752]: > Dec 05 10:37:09.068335 master-0 kubenswrapper[4752]: E1205 10:37:09.068224 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:09.068335 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator_6f76d12f-5406-47e2-8337-2f50e35376d6_0(7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe): error adding pod openshift-apiserver-operator_openshift-apiserver-operator-7bf7f6b755-hdjv7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe" Netns:"/var/run/netns/895de962-b7a8-4b19-b798-ff7111d62b48" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-apiserver-operator;K8S_POD_NAME=openshift-apiserver-operator-7bf7f6b755-hdjv7;K8S_POD_INFRA_CONTAINER_ID=7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe;K8S_POD_UID=6f76d12f-5406-47e2-8337-2f50e35376d6" Path:"" ERRORED: error configuring pod [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7] networking: [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7/6f76d12f-5406-47e2-8337-2f50e35376d6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.068335 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.068335 master-0 kubenswrapper[4752]: > pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:09.068335 master-0 kubenswrapper[4752]: E1205 10:37:09.068252 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:09.068335 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator_6f76d12f-5406-47e2-8337-2f50e35376d6_0(7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe): error adding pod openshift-apiserver-operator_openshift-apiserver-operator-7bf7f6b755-hdjv7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe" Netns:"/var/run/netns/895de962-b7a8-4b19-b798-ff7111d62b48" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-apiserver-operator;K8S_POD_NAME=openshift-apiserver-operator-7bf7f6b755-hdjv7;K8S_POD_INFRA_CONTAINER_ID=7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe;K8S_POD_UID=6f76d12f-5406-47e2-8337-2f50e35376d6" Path:"" ERRORED: error configuring pod [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7] networking: [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7/6f76d12f-5406-47e2-8337-2f50e35376d6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.068335 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.068335 master-0 kubenswrapper[4752]: > pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:09.068617 master-0 kubenswrapper[4752]: E1205 10:37:09.068322 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator(6f76d12f-5406-47e2-8337-2f50e35376d6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator(6f76d12f-5406-47e2-8337-2f50e35376d6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator_6f76d12f-5406-47e2-8337-2f50e35376d6_0(7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe): error adding pod openshift-apiserver-operator_openshift-apiserver-operator-7bf7f6b755-hdjv7 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe\\\" Netns:\\\"/var/run/netns/895de962-b7a8-4b19-b798-ff7111d62b48\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-apiserver-operator;K8S_POD_NAME=openshift-apiserver-operator-7bf7f6b755-hdjv7;K8S_POD_INFRA_CONTAINER_ID=7de46553d19ad30ed71bffa56925d8e509ef195c379ab73448fece3f5873cffe;K8S_POD_UID=6f76d12f-5406-47e2-8337-2f50e35376d6\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7] networking: [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7/6f76d12f-5406-47e2-8337-2f50e35376d6:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" podUID="6f76d12f-5406-47e2-8337-2f50e35376d6" Dec 05 10:37:09.090034 master-0 kubenswrapper[4752]: E1205 10:37:09.089995 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:09.090034 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator_d1c3b7dd-f25e-4983-8a94-084f863fd5b9_0(9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80): error adding pod openshift-service-ca-operator_service-ca-operator-77758bc754-9lzv4 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80" Netns:"/var/run/netns/c81653a0-0ace-4554-9174-59a9f78d7202" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-service-ca-operator;K8S_POD_NAME=service-ca-operator-77758bc754-9lzv4;K8S_POD_INFRA_CONTAINER_ID=9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80;K8S_POD_UID=d1c3b7dd-f25e-4983-8a94-084f863fd5b9" Path:"" ERRORED: error configuring pod [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4] networking: [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4/d1c3b7dd-f25e-4983-8a94-084f863fd5b9:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.090034 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.090034 master-0 kubenswrapper[4752]: > Dec 05 10:37:09.090253 master-0 kubenswrapper[4752]: E1205 10:37:09.090051 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:09.090253 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator_d1c3b7dd-f25e-4983-8a94-084f863fd5b9_0(9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80): error adding pod openshift-service-ca-operator_service-ca-operator-77758bc754-9lzv4 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80" Netns:"/var/run/netns/c81653a0-0ace-4554-9174-59a9f78d7202" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-service-ca-operator;K8S_POD_NAME=service-ca-operator-77758bc754-9lzv4;K8S_POD_INFRA_CONTAINER_ID=9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80;K8S_POD_UID=d1c3b7dd-f25e-4983-8a94-084f863fd5b9" Path:"" ERRORED: error configuring pod [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4] networking: [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4/d1c3b7dd-f25e-4983-8a94-084f863fd5b9:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.090253 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.090253 master-0 kubenswrapper[4752]: > pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:09.090253 master-0 kubenswrapper[4752]: E1205 10:37:09.090074 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:09.090253 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator_d1c3b7dd-f25e-4983-8a94-084f863fd5b9_0(9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80): error adding pod openshift-service-ca-operator_service-ca-operator-77758bc754-9lzv4 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80" Netns:"/var/run/netns/c81653a0-0ace-4554-9174-59a9f78d7202" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-service-ca-operator;K8S_POD_NAME=service-ca-operator-77758bc754-9lzv4;K8S_POD_INFRA_CONTAINER_ID=9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80;K8S_POD_UID=d1c3b7dd-f25e-4983-8a94-084f863fd5b9" Path:"" ERRORED: error configuring pod [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4] networking: [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4/d1c3b7dd-f25e-4983-8a94-084f863fd5b9:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.090253 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.090253 master-0 kubenswrapper[4752]: > pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:09.090557 master-0 kubenswrapper[4752]: E1205 10:37:09.090125 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator(d1c3b7dd-f25e-4983-8a94-084f863fd5b9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator(d1c3b7dd-f25e-4983-8a94-084f863fd5b9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator_d1c3b7dd-f25e-4983-8a94-084f863fd5b9_0(9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80): error adding pod openshift-service-ca-operator_service-ca-operator-77758bc754-9lzv4 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80\\\" Netns:\\\"/var/run/netns/c81653a0-0ace-4554-9174-59a9f78d7202\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-service-ca-operator;K8S_POD_NAME=service-ca-operator-77758bc754-9lzv4;K8S_POD_INFRA_CONTAINER_ID=9a7bd084e540e0c7dd9ce04029553ee995d005de79e4a78e51c5e111411dce80;K8S_POD_UID=d1c3b7dd-f25e-4983-8a94-084f863fd5b9\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4] networking: [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4/d1c3b7dd-f25e-4983-8a94-084f863fd5b9:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" podUID="d1c3b7dd-f25e-4983-8a94-084f863fd5b9" Dec 05 10:37:09.095944 master-0 kubenswrapper[4752]: I1205 10:37:09.095910 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:09.117942 master-0 kubenswrapper[4752]: E1205 10:37:09.117877 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:09.117942 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator_1e69ce9e-4e6f-4015-9ba6-5a7942570190_0(e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa): error adding pod openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa" Netns:"/var/run/netns/1aff68e5-52f4-41ac-8a17-a3a5bbb11ec7" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager-operator;K8S_POD_NAME=openshift-controller-manager-operator-6c8676f99d-cwvk5;K8S_POD_INFRA_CONTAINER_ID=e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa;K8S_POD_UID=1e69ce9e-4e6f-4015-9ba6-5a7942570190" Path:"" ERRORED: error configuring pod [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5] networking: [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5/1e69ce9e-4e6f-4015-9ba6-5a7942570190:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.117942 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.117942 master-0 kubenswrapper[4752]: > Dec 05 10:37:09.118071 master-0 kubenswrapper[4752]: E1205 10:37:09.118002 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:09.118071 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator_1e69ce9e-4e6f-4015-9ba6-5a7942570190_0(e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa): error adding pod openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa" Netns:"/var/run/netns/1aff68e5-52f4-41ac-8a17-a3a5bbb11ec7" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager-operator;K8S_POD_NAME=openshift-controller-manager-operator-6c8676f99d-cwvk5;K8S_POD_INFRA_CONTAINER_ID=e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa;K8S_POD_UID=1e69ce9e-4e6f-4015-9ba6-5a7942570190" Path:"" ERRORED: error configuring pod [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5] networking: [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5/1e69ce9e-4e6f-4015-9ba6-5a7942570190:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.118071 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.118071 master-0 kubenswrapper[4752]: > pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:09.118071 master-0 kubenswrapper[4752]: E1205 10:37:09.118027 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:09.118071 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator_1e69ce9e-4e6f-4015-9ba6-5a7942570190_0(e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa): error adding pod openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa" Netns:"/var/run/netns/1aff68e5-52f4-41ac-8a17-a3a5bbb11ec7" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager-operator;K8S_POD_NAME=openshift-controller-manager-operator-6c8676f99d-cwvk5;K8S_POD_INFRA_CONTAINER_ID=e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa;K8S_POD_UID=1e69ce9e-4e6f-4015-9ba6-5a7942570190" Path:"" ERRORED: error configuring pod [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5] networking: [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5/1e69ce9e-4e6f-4015-9ba6-5a7942570190:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.118071 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.118071 master-0 kubenswrapper[4752]: > pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:09.118546 master-0 kubenswrapper[4752]: E1205 10:37:09.118184 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator(1e69ce9e-4e6f-4015-9ba6-5a7942570190)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator(1e69ce9e-4e6f-4015-9ba6-5a7942570190)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator_1e69ce9e-4e6f-4015-9ba6-5a7942570190_0(e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa): error adding pod openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa\\\" Netns:\\\"/var/run/netns/1aff68e5-52f4-41ac-8a17-a3a5bbb11ec7\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager-operator;K8S_POD_NAME=openshift-controller-manager-operator-6c8676f99d-cwvk5;K8S_POD_INFRA_CONTAINER_ID=e9d724363dc532bee68a938fe6a28d6a7e2f53361c81cf55ea68ee02809791aa;K8S_POD_UID=1e69ce9e-4e6f-4015-9ba6-5a7942570190\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5] networking: [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5/1e69ce9e-4e6f-4015-9ba6-5a7942570190:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" podUID="1e69ce9e-4e6f-4015-9ba6-5a7942570190" Dec 05 10:37:09.136212 master-0 kubenswrapper[4752]: I1205 10:37:09.136161 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:09.136325 master-0 kubenswrapper[4752]: I1205 10:37:09.136212 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:09.136325 master-0 kubenswrapper[4752]: I1205 10:37:09.136245 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:09.136414 master-0 kubenswrapper[4752]: E1205 10:37:09.136350 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: secret "package-server-manager-serving-cert" not found Dec 05 10:37:09.136414 master-0 kubenswrapper[4752]: E1205 10:37:09.136360 4752 secret.go:189] Couldn't get secret openshift-dns-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:09.137282 master-0 kubenswrapper[4752]: E1205 10:37:09.136454 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert podName:8c649a16-c187-412e-b5da-62a00bee38ab nodeName:}" failed. No retries permitted until 2025-12-05 10:37:10.136406198 +0000 UTC m=+111.676529537 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert") pod "package-server-manager-67477646d4-nm8cn" (UID: "8c649a16-c187-412e-b5da-62a00bee38ab") : secret "package-server-manager-serving-cert" not found Dec 05 10:37:09.137282 master-0 kubenswrapper[4752]: E1205 10:37:09.136532 4752 secret.go:189] Couldn't get secret openshift-ingress-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:09.137371 master-0 kubenswrapper[4752]: I1205 10:37:09.137293 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:09.137371 master-0 kubenswrapper[4752]: E1205 10:37:09.137317 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls podName:22676fac-b770-4937-9bee-7478bd1babb7 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:10.137296751 +0000 UTC m=+111.677420080 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls") pod "ingress-operator-8649c48786-cgt5x" (UID: "22676fac-b770-4937-9bee-7478bd1babb7") : secret "metrics-tls" not found Dec 05 10:37:09.137371 master-0 kubenswrapper[4752]: E1205 10:37:09.137343 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls podName:5a1bdc70-6412-47e0-8330-04d796cc8d55 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:10.137324132 +0000 UTC m=+111.677447511 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls") pod "dns-operator-7c56cf9b74-x6t9h" (UID: "5a1bdc70-6412-47e0-8330-04d796cc8d55") : secret "metrics-tls" not found Dec 05 10:37:09.137566 master-0 kubenswrapper[4752]: I1205 10:37:09.137531 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:09.137628 master-0 kubenswrapper[4752]: I1205 10:37:09.137576 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:09.137628 master-0 kubenswrapper[4752]: I1205 10:37:09.137615 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:09.137712 master-0 kubenswrapper[4752]: I1205 10:37:09.137672 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:09.137742 master-0 kubenswrapper[4752]: I1205 10:37:09.137715 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:09.137770 master-0 kubenswrapper[4752]: I1205 10:37:09.137750 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:09.137895 master-0 kubenswrapper[4752]: E1205 10:37:09.137855 4752 secret.go:189] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: secret "marketplace-operator-metrics" not found Dec 05 10:37:09.137953 master-0 kubenswrapper[4752]: E1205 10:37:09.137918 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics podName:926263c4-ec5b-41cb-9c30-0c88f636035f nodeName:}" failed. No retries permitted until 2025-12-05 10:37:10.137902816 +0000 UTC m=+111.678026165 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics") pod "marketplace-operator-f797b99b6-z9qcl" (UID: "926263c4-ec5b-41cb-9c30-0c88f636035f") : secret "marketplace-operator-metrics" not found Dec 05 10:37:09.137953 master-0 kubenswrapper[4752]: E1205 10:37:09.137924 4752 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:09.138013 master-0 kubenswrapper[4752]: E1205 10:37:09.137964 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/node-tuning-operator-tls: secret "node-tuning-operator-tls" not found Dec 05 10:37:09.138013 master-0 kubenswrapper[4752]: E1205 10:37:09.137968 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls podName:a722cda9-29a0-4b7f-8e1d-9a8950ed765a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:10.137955147 +0000 UTC m=+111.678078486 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-7ff994598c-kq8qr" (UID: "a722cda9-29a0-4b7f-8e1d-9a8950ed765a") : secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:09.138086 master-0 kubenswrapper[4752]: E1205 10:37:09.138034 4752 secret.go:189] Couldn't get secret openshift-image-registry/image-registry-operator-tls: secret "image-registry-operator-tls" not found Dec 05 10:37:09.138086 master-0 kubenswrapper[4752]: E1205 10:37:09.138041 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:10.138003009 +0000 UTC m=+111.678126408 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "node-tuning-operator-tls" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "node-tuning-operator-tls" not found Dec 05 10:37:09.138086 master-0 kubenswrapper[4752]: E1205 10:37:09.137867 4752 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: secret "multus-admission-controller-secret" not found Dec 05 10:37:09.138086 master-0 kubenswrapper[4752]: E1205 10:37:09.138069 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls podName:f7b29f89-e42d-4e53-ad14-05efdce933f0 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:10.13806039 +0000 UTC m=+111.678183729 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls") pod "cluster-image-registry-operator-6fb9f88b7-f29mb" (UID: "f7b29f89-e42d-4e53-ad14-05efdce933f0") : secret "image-registry-operator-tls" not found Dec 05 10:37:09.138187 master-0 kubenswrapper[4752]: E1205 10:37:09.138080 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: secret "olm-operator-serving-cert" not found Dec 05 10:37:09.138217 master-0 kubenswrapper[4752]: E1205 10:37:09.138108 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs podName:8d76404b-6d62-4a61-b6f6-0c8073eba198 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:10.138096941 +0000 UTC m=+111.678220370 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs") pod "multus-admission-controller-7dfc5b745f-67rx7" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198") : secret "multus-admission-controller-secret" not found Dec 05 10:37:09.138217 master-0 kubenswrapper[4752]: E1205 10:37:09.138186 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/performance-addon-operator-webhook-cert: secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:09.138273 master-0 kubenswrapper[4752]: E1205 10:37:09.138237 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert podName:46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:10.138209704 +0000 UTC m=+111.678333093 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert") pod "olm-operator-7cd7dbb44c-d25sk" (UID: "46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082") : secret "olm-operator-serving-cert" not found Dec 05 10:37:09.138273 master-0 kubenswrapper[4752]: E1205 10:37:09.138256 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:10.138247355 +0000 UTC m=+111.678370794 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:09.159088 master-0 kubenswrapper[4752]: I1205 10:37:09.158975 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:09.227868 master-0 kubenswrapper[4752]: I1205 10:37:09.227740 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-d6wjk" event={"ID":"283122ba-be1c-4516-bd0f-df41c13c098b","Type":"ContainerStarted","Data":"b44ac27d1b7f531e11c3f4d352d89c4a7679a795ec64417e43c87c5d1e725e90"} Dec 05 10:37:09.231970 master-0 kubenswrapper[4752]: I1205 10:37:09.231937 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" event={"ID":"495ba1ea-f844-43ec-8be7-47e738f5428a","Type":"ContainerStarted","Data":"0f7eef52f441a8502b549c35f33d2f76a407f6fbf4e8ce48e17b60957858833b"} Dec 05 10:37:09.231970 master-0 kubenswrapper[4752]: I1205 10:37:09.231961 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" event={"ID":"495ba1ea-f844-43ec-8be7-47e738f5428a","Type":"ContainerStarted","Data":"b45750dc22b25736719284c0f85981c38a1e92dd897a4968ecbe0f7b0f199edd"} Dec 05 10:37:09.231970 master-0 kubenswrapper[4752]: I1205 10:37:09.231972 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" event={"ID":"495ba1ea-f844-43ec-8be7-47e738f5428a","Type":"ContainerStarted","Data":"0082b9fd574df5f9e1c93733d861b3deaa569a21dcada64c9f285fb4978b2ea9"} Dec 05 10:37:09.232069 master-0 kubenswrapper[4752]: I1205 10:37:09.231982 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" event={"ID":"495ba1ea-f844-43ec-8be7-47e738f5428a","Type":"ContainerStarted","Data":"d3b0c91f26ab8d4c25e128d3baca337a2f0cd0f65138c5ce03ee223d5d0a0329"} Dec 05 10:37:09.232069 master-0 kubenswrapper[4752]: I1205 10:37:09.231991 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" event={"ID":"495ba1ea-f844-43ec-8be7-47e738f5428a","Type":"ContainerStarted","Data":"2034ff8965cd329a1bdbeafb07d2602c622fb8b2b43ccc29a20a537696ed6849"} Dec 05 10:37:09.232069 master-0 kubenswrapper[4752]: I1205 10:37:09.231999 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" event={"ID":"495ba1ea-f844-43ec-8be7-47e738f5428a","Type":"ContainerStarted","Data":"af087d068552ad849510375218150f65ad4adca20fb78424fff8d4e99064422a"} Dec 05 10:37:09.234221 master-0 kubenswrapper[4752]: E1205 10:37:09.234174 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:09.234221 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator_fd58232c-a81a-4aee-8b2c-5ffcdded2e23_0(aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a): error adding pod openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a" Netns:"/var/run/netns/a8de418d-d04f-4026-91aa-045e1dd9df9c" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-storage-version-migrator-operator;K8S_POD_NAME=kube-storage-version-migrator-operator-b9c5dfc78-4gqxr;K8S_POD_INFRA_CONTAINER_ID=aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a;K8S_POD_UID=fd58232c-a81a-4aee-8b2c-5ffcdded2e23" Path:"" ERRORED: error configuring pod [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr] networking: [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr/fd58232c-a81a-4aee-8b2c-5ffcdded2e23:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.234221 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.234221 master-0 kubenswrapper[4752]: > Dec 05 10:37:09.234346 master-0 kubenswrapper[4752]: E1205 10:37:09.234241 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:09.234346 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator_fd58232c-a81a-4aee-8b2c-5ffcdded2e23_0(aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a): error adding pod openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a" Netns:"/var/run/netns/a8de418d-d04f-4026-91aa-045e1dd9df9c" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-storage-version-migrator-operator;K8S_POD_NAME=kube-storage-version-migrator-operator-b9c5dfc78-4gqxr;K8S_POD_INFRA_CONTAINER_ID=aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a;K8S_POD_UID=fd58232c-a81a-4aee-8b2c-5ffcdded2e23" Path:"" ERRORED: error configuring pod [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr] networking: [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr/fd58232c-a81a-4aee-8b2c-5ffcdded2e23:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.234346 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.234346 master-0 kubenswrapper[4752]: > pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:09.234346 master-0 kubenswrapper[4752]: E1205 10:37:09.234260 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:09.234346 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator_fd58232c-a81a-4aee-8b2c-5ffcdded2e23_0(aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a): error adding pod openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a" Netns:"/var/run/netns/a8de418d-d04f-4026-91aa-045e1dd9df9c" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-storage-version-migrator-operator;K8S_POD_NAME=kube-storage-version-migrator-operator-b9c5dfc78-4gqxr;K8S_POD_INFRA_CONTAINER_ID=aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a;K8S_POD_UID=fd58232c-a81a-4aee-8b2c-5ffcdded2e23" Path:"" ERRORED: error configuring pod [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr] networking: [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr/fd58232c-a81a-4aee-8b2c-5ffcdded2e23:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.234346 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.234346 master-0 kubenswrapper[4752]: > pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:09.234554 master-0 kubenswrapper[4752]: E1205 10:37:09.234315 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator(fd58232c-a81a-4aee-8b2c-5ffcdded2e23)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator(fd58232c-a81a-4aee-8b2c-5ffcdded2e23)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator_fd58232c-a81a-4aee-8b2c-5ffcdded2e23_0(aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a): error adding pod openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a\\\" Netns:\\\"/var/run/netns/a8de418d-d04f-4026-91aa-045e1dd9df9c\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-storage-version-migrator-operator;K8S_POD_NAME=kube-storage-version-migrator-operator-b9c5dfc78-4gqxr;K8S_POD_INFRA_CONTAINER_ID=aa0e0781f298f2f2f19273c11a40fb8628849d12fef5c3dd8107e613a898329a;K8S_POD_UID=fd58232c-a81a-4aee-8b2c-5ffcdded2e23\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr] networking: [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr/fd58232c-a81a-4aee-8b2c-5ffcdded2e23:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" podUID="fd58232c-a81a-4aee-8b2c-5ffcdded2e23" Dec 05 10:37:09.344873 master-0 kubenswrapper[4752]: E1205 10:37:09.344821 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:09.344873 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator_87909f47-f2d7-46f8-a1c8-27336cdcce5d_0(8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808): error adding pod openshift-cluster-storage-operator_csi-snapshot-controller-operator-6bc8656fdc-vd94f to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808" Netns:"/var/run/netns/39e5f92a-2ead-43ac-8a4a-2c5dd2df287b" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-storage-operator;K8S_POD_NAME=csi-snapshot-controller-operator-6bc8656fdc-vd94f;K8S_POD_INFRA_CONTAINER_ID=8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808;K8S_POD_UID=87909f47-f2d7-46f8-a1c8-27336cdcce5d" Path:"" ERRORED: error configuring pod [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f] networking: [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f/87909f47-f2d7-46f8-a1c8-27336cdcce5d:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.344873 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.344873 master-0 kubenswrapper[4752]: > Dec 05 10:37:09.345002 master-0 kubenswrapper[4752]: E1205 10:37:09.344911 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:09.345002 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator_87909f47-f2d7-46f8-a1c8-27336cdcce5d_0(8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808): error adding pod openshift-cluster-storage-operator_csi-snapshot-controller-operator-6bc8656fdc-vd94f to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808" Netns:"/var/run/netns/39e5f92a-2ead-43ac-8a4a-2c5dd2df287b" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-storage-operator;K8S_POD_NAME=csi-snapshot-controller-operator-6bc8656fdc-vd94f;K8S_POD_INFRA_CONTAINER_ID=8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808;K8S_POD_UID=87909f47-f2d7-46f8-a1c8-27336cdcce5d" Path:"" ERRORED: error configuring pod [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f] networking: [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f/87909f47-f2d7-46f8-a1c8-27336cdcce5d:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.345002 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.345002 master-0 kubenswrapper[4752]: > pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:09.345002 master-0 kubenswrapper[4752]: E1205 10:37:09.344940 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:09.345002 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator_87909f47-f2d7-46f8-a1c8-27336cdcce5d_0(8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808): error adding pod openshift-cluster-storage-operator_csi-snapshot-controller-operator-6bc8656fdc-vd94f to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808" Netns:"/var/run/netns/39e5f92a-2ead-43ac-8a4a-2c5dd2df287b" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-storage-operator;K8S_POD_NAME=csi-snapshot-controller-operator-6bc8656fdc-vd94f;K8S_POD_INFRA_CONTAINER_ID=8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808;K8S_POD_UID=87909f47-f2d7-46f8-a1c8-27336cdcce5d" Path:"" ERRORED: error configuring pod [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f] networking: [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f/87909f47-f2d7-46f8-a1c8-27336cdcce5d:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:09.345002 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:09.345002 master-0 kubenswrapper[4752]: > pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:09.345189 master-0 kubenswrapper[4752]: E1205 10:37:09.345014 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator(87909f47-f2d7-46f8-a1c8-27336cdcce5d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator(87909f47-f2d7-46f8-a1c8-27336cdcce5d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator_87909f47-f2d7-46f8-a1c8-27336cdcce5d_0(8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808): error adding pod openshift-cluster-storage-operator_csi-snapshot-controller-operator-6bc8656fdc-vd94f to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808\\\" Netns:\\\"/var/run/netns/39e5f92a-2ead-43ac-8a4a-2c5dd2df287b\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-storage-operator;K8S_POD_NAME=csi-snapshot-controller-operator-6bc8656fdc-vd94f;K8S_POD_INFRA_CONTAINER_ID=8a64a1698412099472a519a8b8faaba571514bd0efd8661727625771dae84808;K8S_POD_UID=87909f47-f2d7-46f8-a1c8-27336cdcce5d\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f] networking: [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f/87909f47-f2d7-46f8-a1c8-27336cdcce5d:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" podUID="87909f47-f2d7-46f8-a1c8-27336cdcce5d" Dec 05 10:37:09.795291 master-0 kubenswrapper[4752]: I1205 10:37:09.795195 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:09.795291 master-0 kubenswrapper[4752]: I1205 10:37:09.795264 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:09.797158 master-0 kubenswrapper[4752]: I1205 10:37:09.797086 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 10:37:09.797158 master-0 kubenswrapper[4752]: I1205 10:37:09.797137 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 10:37:09.798501 master-0 kubenswrapper[4752]: I1205 10:37:09.798476 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 10:37:10.148649 master-0 kubenswrapper[4752]: I1205 10:37:10.148591 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:10.148816 master-0 kubenswrapper[4752]: E1205 10:37:10.148775 4752 secret.go:189] Couldn't get secret openshift-ingress-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:10.148875 master-0 kubenswrapper[4752]: E1205 10:37:10.148833 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls podName:22676fac-b770-4937-9bee-7478bd1babb7 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:12.148815745 +0000 UTC m=+113.688939154 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls") pod "ingress-operator-8649c48786-cgt5x" (UID: "22676fac-b770-4937-9bee-7478bd1babb7") : secret "metrics-tls" not found Dec 05 10:37:10.148875 master-0 kubenswrapper[4752]: I1205 10:37:10.148838 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:10.149022 master-0 kubenswrapper[4752]: E1205 10:37:10.148986 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: secret "olm-operator-serving-cert" not found Dec 05 10:37:10.149059 master-0 kubenswrapper[4752]: I1205 10:37:10.148986 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:10.149088 master-0 kubenswrapper[4752]: E1205 10:37:10.149056 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/performance-addon-operator-webhook-cert: secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:10.149088 master-0 kubenswrapper[4752]: E1205 10:37:10.149069 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert podName:46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:12.14904818 +0000 UTC m=+113.689171609 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert") pod "olm-operator-7cd7dbb44c-d25sk" (UID: "46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082") : secret "olm-operator-serving-cert" not found Dec 05 10:37:10.149140 master-0 kubenswrapper[4752]: I1205 10:37:10.149101 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:10.149183 master-0 kubenswrapper[4752]: I1205 10:37:10.149166 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:10.149213 master-0 kubenswrapper[4752]: E1205 10:37:10.149178 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:12.149166263 +0000 UTC m=+113.689289692 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:10.149243 master-0 kubenswrapper[4752]: E1205 10:37:10.149227 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/node-tuning-operator-tls: secret "node-tuning-operator-tls" not found Dec 05 10:37:10.149270 master-0 kubenswrapper[4752]: E1205 10:37:10.149258 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:12.149249035 +0000 UTC m=+113.689372484 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "node-tuning-operator-tls" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "node-tuning-operator-tls" not found Dec 05 10:37:10.149304 master-0 kubenswrapper[4752]: E1205 10:37:10.149279 4752 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: secret "multus-admission-controller-secret" not found Dec 05 10:37:10.149337 master-0 kubenswrapper[4752]: I1205 10:37:10.149302 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:10.149337 master-0 kubenswrapper[4752]: E1205 10:37:10.149333 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs podName:8d76404b-6d62-4a61-b6f6-0c8073eba198 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:12.149316367 +0000 UTC m=+113.689439716 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs") pod "multus-admission-controller-7dfc5b745f-67rx7" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198") : secret "multus-admission-controller-secret" not found Dec 05 10:37:10.149389 master-0 kubenswrapper[4752]: I1205 10:37:10.149355 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:10.149389 master-0 kubenswrapper[4752]: I1205 10:37:10.149385 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:10.149474 master-0 kubenswrapper[4752]: E1205 10:37:10.149396 4752 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:10.149474 master-0 kubenswrapper[4752]: I1205 10:37:10.149407 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:10.149474 master-0 kubenswrapper[4752]: I1205 10:37:10.149450 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:10.149637 master-0 kubenswrapper[4752]: E1205 10:37:10.149479 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls podName:a722cda9-29a0-4b7f-8e1d-9a8950ed765a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:12.149459682 +0000 UTC m=+113.689583101 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-7ff994598c-kq8qr" (UID: "a722cda9-29a0-4b7f-8e1d-9a8950ed765a") : secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:10.149637 master-0 kubenswrapper[4752]: E1205 10:37:10.149532 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: secret "package-server-manager-serving-cert" not found Dec 05 10:37:10.149637 master-0 kubenswrapper[4752]: E1205 10:37:10.149604 4752 secret.go:189] Couldn't get secret openshift-image-registry/image-registry-operator-tls: secret "image-registry-operator-tls" not found Dec 05 10:37:10.149718 master-0 kubenswrapper[4752]: E1205 10:37:10.149637 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls podName:f7b29f89-e42d-4e53-ad14-05efdce933f0 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:12.149626886 +0000 UTC m=+113.689750225 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls") pod "cluster-image-registry-operator-6fb9f88b7-f29mb" (UID: "f7b29f89-e42d-4e53-ad14-05efdce933f0") : secret "image-registry-operator-tls" not found Dec 05 10:37:10.149718 master-0 kubenswrapper[4752]: E1205 10:37:10.149534 4752 secret.go:189] Couldn't get secret openshift-dns-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:10.149718 master-0 kubenswrapper[4752]: E1205 10:37:10.149574 4752 secret.go:189] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: secret "marketplace-operator-metrics" not found Dec 05 10:37:10.149817 master-0 kubenswrapper[4752]: E1205 10:37:10.149669 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert podName:8c649a16-c187-412e-b5da-62a00bee38ab nodeName:}" failed. No retries permitted until 2025-12-05 10:37:12.149653056 +0000 UTC m=+113.689776475 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert") pod "package-server-manager-67477646d4-nm8cn" (UID: "8c649a16-c187-412e-b5da-62a00bee38ab") : secret "package-server-manager-serving-cert" not found Dec 05 10:37:10.149868 master-0 kubenswrapper[4752]: E1205 10:37:10.149822 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls podName:5a1bdc70-6412-47e0-8330-04d796cc8d55 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:12.14980835 +0000 UTC m=+113.689931749 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls") pod "dns-operator-7c56cf9b74-x6t9h" (UID: "5a1bdc70-6412-47e0-8330-04d796cc8d55") : secret "metrics-tls" not found Dec 05 10:37:10.149868 master-0 kubenswrapper[4752]: E1205 10:37:10.149840 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics podName:926263c4-ec5b-41cb-9c30-0c88f636035f nodeName:}" failed. No retries permitted until 2025-12-05 10:37:12.149831101 +0000 UTC m=+113.689954570 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics") pod "marketplace-operator-f797b99b6-z9qcl" (UID: "926263c4-ec5b-41cb-9c30-0c88f636035f") : secret "marketplace-operator-metrics" not found Dec 05 10:37:11.242436 master-0 kubenswrapper[4752]: I1205 10:37:11.242284 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" event={"ID":"495ba1ea-f844-43ec-8be7-47e738f5428a","Type":"ContainerStarted","Data":"7a3716291cc973e7a7cd085a12bbf05bdaa8f982556c203be5b35bc4e9c2e8cc"} Dec 05 10:37:12.171629 master-0 kubenswrapper[4752]: I1205 10:37:12.171347 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:12.171629 master-0 kubenswrapper[4752]: I1205 10:37:12.171471 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:12.171629 master-0 kubenswrapper[4752]: I1205 10:37:12.171512 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:12.171629 master-0 kubenswrapper[4752]: I1205 10:37:12.171566 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:12.171629 master-0 kubenswrapper[4752]: I1205 10:37:12.171610 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:12.171629 master-0 kubenswrapper[4752]: I1205 10:37:12.171646 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.171674 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/node-tuning-operator-tls: secret "node-tuning-operator-tls" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.171701 4752 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.171757 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:16.17173348 +0000 UTC m=+117.711856829 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "node-tuning-operator-tls" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "node-tuning-operator-tls" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.171798 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls podName:a722cda9-29a0-4b7f-8e1d-9a8950ed765a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:16.171768991 +0000 UTC m=+117.711892360 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-7ff994598c-kq8qr" (UID: "a722cda9-29a0-4b7f-8e1d-9a8950ed765a") : secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.171815 4752 secret.go:189] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: secret "marketplace-operator-metrics" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.171937 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics podName:926263c4-ec5b-41cb-9c30-0c88f636035f nodeName:}" failed. No retries permitted until 2025-12-05 10:37:16.171912255 +0000 UTC m=+117.712035684 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics") pod "marketplace-operator-f797b99b6-z9qcl" (UID: "926263c4-ec5b-41cb-9c30-0c88f636035f") : secret "marketplace-operator-metrics" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.171945 4752 secret.go:189] Couldn't get secret openshift-image-registry/image-registry-operator-tls: secret "image-registry-operator-tls" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.172022 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: secret "package-server-manager-serving-cert" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.172043 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls podName:f7b29f89-e42d-4e53-ad14-05efdce933f0 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:16.172023137 +0000 UTC m=+117.712146466 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls") pod "cluster-image-registry-operator-6fb9f88b7-f29mb" (UID: "f7b29f89-e42d-4e53-ad14-05efdce933f0") : secret "image-registry-operator-tls" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.172069 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert podName:8c649a16-c187-412e-b5da-62a00bee38ab nodeName:}" failed. No retries permitted until 2025-12-05 10:37:16.172055548 +0000 UTC m=+117.712178987 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert") pod "package-server-manager-67477646d4-nm8cn" (UID: "8c649a16-c187-412e-b5da-62a00bee38ab") : secret "package-server-manager-serving-cert" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.172115 4752 secret.go:189] Couldn't get secret openshift-dns-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.172153 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls podName:5a1bdc70-6412-47e0-8330-04d796cc8d55 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:16.17214301 +0000 UTC m=+117.712266509 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls") pod "dns-operator-7c56cf9b74-x6t9h" (UID: "5a1bdc70-6412-47e0-8330-04d796cc8d55") : secret "metrics-tls" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: I1205 10:37:12.172114 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.172182 4752 secret.go:189] Couldn't get secret openshift-ingress-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:12.172268 master-0 kubenswrapper[4752]: E1205 10:37:12.172218 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls podName:22676fac-b770-4937-9bee-7478bd1babb7 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:16.172208562 +0000 UTC m=+117.712331901 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls") pod "ingress-operator-8649c48786-cgt5x" (UID: "22676fac-b770-4937-9bee-7478bd1babb7") : secret "metrics-tls" not found Dec 05 10:37:12.173250 master-0 kubenswrapper[4752]: I1205 10:37:12.172240 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:12.173250 master-0 kubenswrapper[4752]: I1205 10:37:12.172280 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:12.173250 master-0 kubenswrapper[4752]: E1205 10:37:12.172369 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/performance-addon-operator-webhook-cert: secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:12.173250 master-0 kubenswrapper[4752]: E1205 10:37:12.172409 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:16.172396377 +0000 UTC m=+117.712519856 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:12.173250 master-0 kubenswrapper[4752]: I1205 10:37:12.172466 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:12.173250 master-0 kubenswrapper[4752]: E1205 10:37:12.172571 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: secret "olm-operator-serving-cert" not found Dec 05 10:37:12.173250 master-0 kubenswrapper[4752]: E1205 10:37:12.172608 4752 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: secret "multus-admission-controller-secret" not found Dec 05 10:37:12.173250 master-0 kubenswrapper[4752]: E1205 10:37:12.172641 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert podName:46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:16.172618882 +0000 UTC m=+117.712742251 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert") pod "olm-operator-7cd7dbb44c-d25sk" (UID: "46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082") : secret "olm-operator-serving-cert" not found Dec 05 10:37:12.173250 master-0 kubenswrapper[4752]: E1205 10:37:12.172675 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs podName:8d76404b-6d62-4a61-b6f6-0c8073eba198 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:16.172660763 +0000 UTC m=+117.712784132 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs") pod "multus-admission-controller-7dfc5b745f-67rx7" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198") : secret "multus-admission-controller-secret" not found Dec 05 10:37:13.256772 master-0 kubenswrapper[4752]: I1205 10:37:13.256699 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" event={"ID":"495ba1ea-f844-43ec-8be7-47e738f5428a","Type":"ContainerStarted","Data":"896ccf421e9504196fb982f341ca8f6bc0493a0e5df7567cd630846af2a1529d"} Dec 05 10:37:13.257678 master-0 kubenswrapper[4752]: I1205 10:37:13.257068 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:13.257678 master-0 kubenswrapper[4752]: I1205 10:37:13.257240 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:13.257678 master-0 kubenswrapper[4752]: I1205 10:37:13.257280 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:13.281386 master-0 kubenswrapper[4752]: I1205 10:37:13.281333 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:13.282446 master-0 kubenswrapper[4752]: I1205 10:37:13.282394 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:13.311332 master-0 kubenswrapper[4752]: I1205 10:37:13.311272 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" podStartSLOduration=7.311224942 podStartE2EDuration="7.311224942s" podCreationTimestamp="2025-12-05 10:37:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:37:13.286548512 +0000 UTC m=+114.826671861" watchObservedRunningTime="2025-12-05 10:37:13.311224942 +0000 UTC m=+114.851348271" Dec 05 10:37:13.986299 master-0 kubenswrapper[4752]: I1205 10:37:13.985955 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl"] Dec 05 10:37:13.986698 master-0 kubenswrapper[4752]: I1205 10:37:13.986374 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:13.986783 master-0 kubenswrapper[4752]: I1205 10:37:13.986769 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:13.991321 master-0 kubenswrapper[4752]: I1205 10:37:13.991277 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn"] Dec 05 10:37:13.993647 master-0 kubenswrapper[4752]: I1205 10:37:13.993593 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c"] Dec 05 10:37:13.993776 master-0 kubenswrapper[4752]: I1205 10:37:13.993763 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:13.994191 master-0 kubenswrapper[4752]: I1205 10:37:13.994155 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:13.994319 master-0 kubenswrapper[4752]: I1205 10:37:13.994283 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb"] Dec 05 10:37:13.999457 master-0 kubenswrapper[4752]: I1205 10:37:13.996631 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4"] Dec 05 10:37:13.999457 master-0 kubenswrapper[4752]: I1205 10:37:13.996749 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:13.999457 master-0 kubenswrapper[4752]: I1205 10:37:13.997084 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:13.999457 master-0 kubenswrapper[4752]: I1205 10:37:13.997268 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr"] Dec 05 10:37:14.008790 master-0 kubenswrapper[4752]: I1205 10:37:14.006927 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv"] Dec 05 10:37:14.008790 master-0 kubenswrapper[4752]: I1205 10:37:14.007106 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:14.008790 master-0 kubenswrapper[4752]: I1205 10:37:14.007544 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:14.009606 master-0 kubenswrapper[4752]: I1205 10:37:14.009415 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h"] Dec 05 10:37:14.012522 master-0 kubenswrapper[4752]: I1205 10:37:14.011215 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-8649c48786-cgt5x"] Dec 05 10:37:14.012522 master-0 kubenswrapper[4752]: I1205 10:37:14.012382 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t"] Dec 05 10:37:14.012522 master-0 kubenswrapper[4752]: I1205 10:37:14.012485 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:14.012805 master-0 kubenswrapper[4752]: I1205 10:37:14.012791 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:14.014239 master-0 kubenswrapper[4752]: I1205 10:37:14.014202 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7"] Dec 05 10:37:14.014390 master-0 kubenswrapper[4752]: I1205 10:37:14.014372 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:14.014849 master-0 kubenswrapper[4752]: I1205 10:37:14.014787 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:14.015377 master-0 kubenswrapper[4752]: I1205 10:37:14.015343 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-7dfc5b745f-67rx7"] Dec 05 10:37:14.016499 master-0 kubenswrapper[4752]: I1205 10:37:14.016472 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk"] Dec 05 10:37:14.016991 master-0 kubenswrapper[4752]: I1205 10:37:14.016965 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5"] Dec 05 10:37:14.017107 master-0 kubenswrapper[4752]: I1205 10:37:14.017070 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:14.017385 master-0 kubenswrapper[4752]: I1205 10:37:14.017367 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:14.020558 master-0 kubenswrapper[4752]: I1205 10:37:14.020532 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr"] Dec 05 10:37:14.020653 master-0 kubenswrapper[4752]: I1205 10:37:14.020627 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:14.020916 master-0 kubenswrapper[4752]: I1205 10:37:14.020899 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:14.023962 master-0 kubenswrapper[4752]: I1205 10:37:14.023560 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f"] Dec 05 10:37:14.023962 master-0 kubenswrapper[4752]: I1205 10:37:14.023926 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:14.024295 master-0 kubenswrapper[4752]: I1205 10:37:14.024275 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:14.026480 master-0 kubenswrapper[4752]: I1205 10:37:14.025103 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-f797b99b6-z9qcl"] Dec 05 10:37:14.026771 master-0 kubenswrapper[4752]: I1205 10:37:14.026705 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p"] Dec 05 10:37:14.027073 master-0 kubenswrapper[4752]: I1205 10:37:14.027039 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27"] Dec 05 10:37:14.027173 master-0 kubenswrapper[4752]: I1205 10:37:14.027140 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:14.027484 master-0 kubenswrapper[4752]: I1205 10:37:14.027461 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:14.028010 master-0 kubenswrapper[4752]: I1205 10:37:14.027977 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8"] Dec 05 10:37:14.028134 master-0 kubenswrapper[4752]: I1205 10:37:14.028099 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:14.029166 master-0 kubenswrapper[4752]: I1205 10:37:14.028605 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:14.038452 master-0 kubenswrapper[4752]: I1205 10:37:14.035043 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8"] Dec 05 10:37:14.038452 master-0 kubenswrapper[4752]: I1205 10:37:14.035184 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:14.038452 master-0 kubenswrapper[4752]: I1205 10:37:14.035576 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:14.198897 master-0 kubenswrapper[4752]: E1205 10:37:14.198847 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.198897 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator_444f8808-e454-4015-9e20-429e715a08c7_0(50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97): error adding pod openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97" Netns:"/var/run/netns/36ce2eb9-0b64-4b0c-b9bb-54be4d998e79" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-apiserver-operator;K8S_POD_NAME=kube-apiserver-operator-765d9ff747-p57fl;K8S_POD_INFRA_CONTAINER_ID=50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97;K8S_POD_UID=444f8808-e454-4015-9e20-429e715a08c7" Path:"" ERRORED: error configuring pod [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl] networking: [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl/444f8808-e454-4015-9e20-429e715a08c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.198897 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.198897 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.199225 master-0 kubenswrapper[4752]: E1205 10:37:14.198933 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.199225 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator_444f8808-e454-4015-9e20-429e715a08c7_0(50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97): error adding pod openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97" Netns:"/var/run/netns/36ce2eb9-0b64-4b0c-b9bb-54be4d998e79" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-apiserver-operator;K8S_POD_NAME=kube-apiserver-operator-765d9ff747-p57fl;K8S_POD_INFRA_CONTAINER_ID=50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97;K8S_POD_UID=444f8808-e454-4015-9e20-429e715a08c7" Path:"" ERRORED: error configuring pod [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl] networking: [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl/444f8808-e454-4015-9e20-429e715a08c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.199225 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.199225 master-0 kubenswrapper[4752]: > pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:14.199225 master-0 kubenswrapper[4752]: E1205 10:37:14.198960 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.199225 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator_444f8808-e454-4015-9e20-429e715a08c7_0(50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97): error adding pod openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97" Netns:"/var/run/netns/36ce2eb9-0b64-4b0c-b9bb-54be4d998e79" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-apiserver-operator;K8S_POD_NAME=kube-apiserver-operator-765d9ff747-p57fl;K8S_POD_INFRA_CONTAINER_ID=50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97;K8S_POD_UID=444f8808-e454-4015-9e20-429e715a08c7" Path:"" ERRORED: error configuring pod [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl] networking: [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl/444f8808-e454-4015-9e20-429e715a08c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.199225 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.199225 master-0 kubenswrapper[4752]: > pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:14.199572 master-0 kubenswrapper[4752]: E1205 10:37:14.199033 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator(444f8808-e454-4015-9e20-429e715a08c7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator(444f8808-e454-4015-9e20-429e715a08c7)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-apiserver-operator-765d9ff747-p57fl_openshift-kube-apiserver-operator_444f8808-e454-4015-9e20-429e715a08c7_0(50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97): error adding pod openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97\\\" Netns:\\\"/var/run/netns/36ce2eb9-0b64-4b0c-b9bb-54be4d998e79\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-apiserver-operator;K8S_POD_NAME=kube-apiserver-operator-765d9ff747-p57fl;K8S_POD_INFRA_CONTAINER_ID=50b9e908c504df062d661ebea36016ad2ba1925c6aaa0df2e2004008f3649d97;K8S_POD_UID=444f8808-e454-4015-9e20-429e715a08c7\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl] networking: [openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl/444f8808-e454-4015-9e20-429e715a08c7:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" podUID="444f8808-e454-4015-9e20-429e715a08c7" Dec 05 10:37:14.213612 master-0 kubenswrapper[4752]: E1205 10:37:14.213555 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.213612 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator_c22d947f-a5b6-4f24-b142-dd201c46293b_0(fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897): error adding pod openshift-config-operator_openshift-config-operator-68758cbcdb-dnpcv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897" Netns:"/var/run/netns/d290f384-3413-41e7-9832-77cc6cd9a00d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-config-operator;K8S_POD_NAME=openshift-config-operator-68758cbcdb-dnpcv;K8S_POD_INFRA_CONTAINER_ID=fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897;K8S_POD_UID=c22d947f-a5b6-4f24-b142-dd201c46293b" Path:"" ERRORED: error configuring pod [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv] networking: [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv/c22d947f-a5b6-4f24-b142-dd201c46293b:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.213612 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.213612 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.213938 master-0 kubenswrapper[4752]: E1205 10:37:14.213626 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.213938 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator_c22d947f-a5b6-4f24-b142-dd201c46293b_0(fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897): error adding pod openshift-config-operator_openshift-config-operator-68758cbcdb-dnpcv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897" Netns:"/var/run/netns/d290f384-3413-41e7-9832-77cc6cd9a00d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-config-operator;K8S_POD_NAME=openshift-config-operator-68758cbcdb-dnpcv;K8S_POD_INFRA_CONTAINER_ID=fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897;K8S_POD_UID=c22d947f-a5b6-4f24-b142-dd201c46293b" Path:"" ERRORED: error configuring pod [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv] networking: [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv/c22d947f-a5b6-4f24-b142-dd201c46293b:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.213938 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.213938 master-0 kubenswrapper[4752]: > pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:14.213938 master-0 kubenswrapper[4752]: E1205 10:37:14.213646 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.213938 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator_c22d947f-a5b6-4f24-b142-dd201c46293b_0(fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897): error adding pod openshift-config-operator_openshift-config-operator-68758cbcdb-dnpcv to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897" Netns:"/var/run/netns/d290f384-3413-41e7-9832-77cc6cd9a00d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-config-operator;K8S_POD_NAME=openshift-config-operator-68758cbcdb-dnpcv;K8S_POD_INFRA_CONTAINER_ID=fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897;K8S_POD_UID=c22d947f-a5b6-4f24-b142-dd201c46293b" Path:"" ERRORED: error configuring pod [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv] networking: [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv/c22d947f-a5b6-4f24-b142-dd201c46293b:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.213938 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.213938 master-0 kubenswrapper[4752]: > pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:14.214191 master-0 kubenswrapper[4752]: E1205 10:37:14.213703 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator(c22d947f-a5b6-4f24-b142-dd201c46293b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator(c22d947f-a5b6-4f24-b142-dd201c46293b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-config-operator-68758cbcdb-dnpcv_openshift-config-operator_c22d947f-a5b6-4f24-b142-dd201c46293b_0(fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897): error adding pod openshift-config-operator_openshift-config-operator-68758cbcdb-dnpcv to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897\\\" Netns:\\\"/var/run/netns/d290f384-3413-41e7-9832-77cc6cd9a00d\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-config-operator;K8S_POD_NAME=openshift-config-operator-68758cbcdb-dnpcv;K8S_POD_INFRA_CONTAINER_ID=fb77cef39bd12026cb1bcb6e3ea3241aca08c9633f69218f3caff98eef141897;K8S_POD_UID=c22d947f-a5b6-4f24-b142-dd201c46293b\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv] networking: [openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv/c22d947f-a5b6-4f24-b142-dd201c46293b:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" podUID="c22d947f-a5b6-4f24-b142-dd201c46293b" Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: E1205 10:37:14.225465 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator_d1c3b7dd-f25e-4983-8a94-084f863fd5b9_0(f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52): error adding pod openshift-service-ca-operator_service-ca-operator-77758bc754-9lzv4 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52" Netns:"/var/run/netns/46efd3e3-8213-4e9f-b467-5098940cbfae" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-service-ca-operator;K8S_POD_NAME=service-ca-operator-77758bc754-9lzv4;K8S_POD_INFRA_CONTAINER_ID=f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52;K8S_POD_UID=d1c3b7dd-f25e-4983-8a94-084f863fd5b9" Path:"" ERRORED: error configuring pod [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4] networking: [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4/d1c3b7dd-f25e-4983-8a94-084f863fd5b9:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: E1205 10:37:14.225520 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator_d1c3b7dd-f25e-4983-8a94-084f863fd5b9_0(f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52): error adding pod openshift-service-ca-operator_service-ca-operator-77758bc754-9lzv4 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52" Netns:"/var/run/netns/46efd3e3-8213-4e9f-b467-5098940cbfae" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-service-ca-operator;K8S_POD_NAME=service-ca-operator-77758bc754-9lzv4;K8S_POD_INFRA_CONTAINER_ID=f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52;K8S_POD_UID=d1c3b7dd-f25e-4983-8a94-084f863fd5b9" Path:"" ERRORED: error configuring pod [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4] networking: [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4/d1c3b7dd-f25e-4983-8a94-084f863fd5b9:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: > pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: E1205 10:37:14.225541 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator_d1c3b7dd-f25e-4983-8a94-084f863fd5b9_0(f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52): error adding pod openshift-service-ca-operator_service-ca-operator-77758bc754-9lzv4 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52" Netns:"/var/run/netns/46efd3e3-8213-4e9f-b467-5098940cbfae" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-service-ca-operator;K8S_POD_NAME=service-ca-operator-77758bc754-9lzv4;K8S_POD_INFRA_CONTAINER_ID=f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52;K8S_POD_UID=d1c3b7dd-f25e-4983-8a94-084f863fd5b9" Path:"" ERRORED: error configuring pod [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4] networking: [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4/d1c3b7dd-f25e-4983-8a94-084f863fd5b9:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.225723 master-0 kubenswrapper[4752]: > pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:14.226264 master-0 kubenswrapper[4752]: E1205 10:37:14.225590 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator(d1c3b7dd-f25e-4983-8a94-084f863fd5b9)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator(d1c3b7dd-f25e-4983-8a94-084f863fd5b9)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_service-ca-operator-77758bc754-9lzv4_openshift-service-ca-operator_d1c3b7dd-f25e-4983-8a94-084f863fd5b9_0(f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52): error adding pod openshift-service-ca-operator_service-ca-operator-77758bc754-9lzv4 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52\\\" Netns:\\\"/var/run/netns/46efd3e3-8213-4e9f-b467-5098940cbfae\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-service-ca-operator;K8S_POD_NAME=service-ca-operator-77758bc754-9lzv4;K8S_POD_INFRA_CONTAINER_ID=f74fc19e206caf778f95d1af81515808edcb225c642d8982035eb3ca5f8d4f52;K8S_POD_UID=d1c3b7dd-f25e-4983-8a94-084f863fd5b9\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4] networking: [openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4/d1c3b7dd-f25e-4983-8a94-084f863fd5b9:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" podUID="d1c3b7dd-f25e-4983-8a94-084f863fd5b9" Dec 05 10:37:14.246765 master-0 kubenswrapper[4752]: E1205 10:37:14.243766 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.246765 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator_fd58232c-a81a-4aee-8b2c-5ffcdded2e23_0(e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49): error adding pod openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49" Netns:"/var/run/netns/354ad038-a651-443d-a996-050c7de75f7d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-storage-version-migrator-operator;K8S_POD_NAME=kube-storage-version-migrator-operator-b9c5dfc78-4gqxr;K8S_POD_INFRA_CONTAINER_ID=e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49;K8S_POD_UID=fd58232c-a81a-4aee-8b2c-5ffcdded2e23" Path:"" ERRORED: error configuring pod [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr] networking: [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr/fd58232c-a81a-4aee-8b2c-5ffcdded2e23:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.246765 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.246765 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.246765 master-0 kubenswrapper[4752]: E1205 10:37:14.243855 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.246765 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator_fd58232c-a81a-4aee-8b2c-5ffcdded2e23_0(e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49): error adding pod openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49" Netns:"/var/run/netns/354ad038-a651-443d-a996-050c7de75f7d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-storage-version-migrator-operator;K8S_POD_NAME=kube-storage-version-migrator-operator-b9c5dfc78-4gqxr;K8S_POD_INFRA_CONTAINER_ID=e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49;K8S_POD_UID=fd58232c-a81a-4aee-8b2c-5ffcdded2e23" Path:"" ERRORED: error configuring pod [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr] networking: [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr/fd58232c-a81a-4aee-8b2c-5ffcdded2e23:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.246765 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.246765 master-0 kubenswrapper[4752]: > pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:14.246765 master-0 kubenswrapper[4752]: E1205 10:37:14.243882 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.246765 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator_fd58232c-a81a-4aee-8b2c-5ffcdded2e23_0(e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49): error adding pod openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49" Netns:"/var/run/netns/354ad038-a651-443d-a996-050c7de75f7d" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-storage-version-migrator-operator;K8S_POD_NAME=kube-storage-version-migrator-operator-b9c5dfc78-4gqxr;K8S_POD_INFRA_CONTAINER_ID=e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49;K8S_POD_UID=fd58232c-a81a-4aee-8b2c-5ffcdded2e23" Path:"" ERRORED: error configuring pod [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr] networking: [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr/fd58232c-a81a-4aee-8b2c-5ffcdded2e23:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.246765 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.247121 master-0 kubenswrapper[4752]: > pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:14.247121 master-0 kubenswrapper[4752]: E1205 10:37:14.243958 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator(fd58232c-a81a-4aee-8b2c-5ffcdded2e23)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator(fd58232c-a81a-4aee-8b2c-5ffcdded2e23)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr_openshift-kube-storage-version-migrator-operator_fd58232c-a81a-4aee-8b2c-5ffcdded2e23_0(e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49): error adding pod openshift-kube-storage-version-migrator-operator_kube-storage-version-migrator-operator-b9c5dfc78-4gqxr to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49\\\" Netns:\\\"/var/run/netns/354ad038-a651-443d-a996-050c7de75f7d\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-storage-version-migrator-operator;K8S_POD_NAME=kube-storage-version-migrator-operator-b9c5dfc78-4gqxr;K8S_POD_INFRA_CONTAINER_ID=e3727575f9c0c969c2e68ee2992d206defd05f3a41c78d6e11f1b751bf6a0b49;K8S_POD_UID=fd58232c-a81a-4aee-8b2c-5ffcdded2e23\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr] networking: [openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr/fd58232c-a81a-4aee-8b2c-5ffcdded2e23:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" podUID="fd58232c-a81a-4aee-8b2c-5ffcdded2e23" Dec 05 10:37:14.251343 master-0 kubenswrapper[4752]: E1205 10:37:14.251293 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.251343 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator_f7a08359-0379-4364-8b0c-ddb58ff605f4_0(0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c): error adding pod openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c" Netns:"/var/run/netns/036aeffb-a53b-4eaf-8c7f-0eb450ff3b27" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-etcd-operator;K8S_POD_NAME=etcd-operator-5bf4d88c6f-n8t5c;K8S_POD_INFRA_CONTAINER_ID=0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c;K8S_POD_UID=f7a08359-0379-4364-8b0c-ddb58ff605f4" Path:"" ERRORED: error configuring pod [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c] networking: [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c/f7a08359-0379-4364-8b0c-ddb58ff605f4:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.251343 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.251343 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.251573 master-0 kubenswrapper[4752]: E1205 10:37:14.251364 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.251573 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator_f7a08359-0379-4364-8b0c-ddb58ff605f4_0(0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c): error adding pod openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c" Netns:"/var/run/netns/036aeffb-a53b-4eaf-8c7f-0eb450ff3b27" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-etcd-operator;K8S_POD_NAME=etcd-operator-5bf4d88c6f-n8t5c;K8S_POD_INFRA_CONTAINER_ID=0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c;K8S_POD_UID=f7a08359-0379-4364-8b0c-ddb58ff605f4" Path:"" ERRORED: error configuring pod [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c] networking: [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c/f7a08359-0379-4364-8b0c-ddb58ff605f4:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.251573 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.251573 master-0 kubenswrapper[4752]: > pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:14.251573 master-0 kubenswrapper[4752]: E1205 10:37:14.251393 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.251573 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator_f7a08359-0379-4364-8b0c-ddb58ff605f4_0(0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c): error adding pod openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c" Netns:"/var/run/netns/036aeffb-a53b-4eaf-8c7f-0eb450ff3b27" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-etcd-operator;K8S_POD_NAME=etcd-operator-5bf4d88c6f-n8t5c;K8S_POD_INFRA_CONTAINER_ID=0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c;K8S_POD_UID=f7a08359-0379-4364-8b0c-ddb58ff605f4" Path:"" ERRORED: error configuring pod [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c] networking: [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c/f7a08359-0379-4364-8b0c-ddb58ff605f4:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.251573 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.251573 master-0 kubenswrapper[4752]: > pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:14.251573 master-0 kubenswrapper[4752]: E1205 10:37:14.251461 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator(f7a08359-0379-4364-8b0c-ddb58ff605f4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator(f7a08359-0379-4364-8b0c-ddb58ff605f4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_etcd-operator-5bf4d88c6f-n8t5c_openshift-etcd-operator_f7a08359-0379-4364-8b0c-ddb58ff605f4_0(0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c): error adding pod openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c\\\" Netns:\\\"/var/run/netns/036aeffb-a53b-4eaf-8c7f-0eb450ff3b27\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-etcd-operator;K8S_POD_NAME=etcd-operator-5bf4d88c6f-n8t5c;K8S_POD_INFRA_CONTAINER_ID=0c5bdb377673d87d670ff147b5fadcd36bc9d0da1081d5ef273c151383f9125c;K8S_POD_UID=f7a08359-0379-4364-8b0c-ddb58ff605f4\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c] networking: [openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c/f7a08359-0379-4364-8b0c-ddb58ff605f4:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" podUID="f7a08359-0379-4364-8b0c-ddb58ff605f4" Dec 05 10:37:14.258224 master-0 kubenswrapper[4752]: E1205 10:37:14.258165 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.258224 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator_11f563d5-89bb-433c-956a-6d5d2492e8f1_0(246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada): error adding pod openshift-kube-controller-manager-operator_kube-controller-manager-operator-848f645654-rmdb8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada" Netns:"/var/run/netns/df186742-b74d-4a28-bd17-a8e51d863583" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager-operator;K8S_POD_NAME=kube-controller-manager-operator-848f645654-rmdb8;K8S_POD_INFRA_CONTAINER_ID=246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada;K8S_POD_UID=11f563d5-89bb-433c-956a-6d5d2492e8f1" Path:"" ERRORED: error configuring pod [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8] networking: [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8/11f563d5-89bb-433c-956a-6d5d2492e8f1:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.258224 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.258224 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.258861 master-0 kubenswrapper[4752]: E1205 10:37:14.258226 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.258861 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator_11f563d5-89bb-433c-956a-6d5d2492e8f1_0(246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada): error adding pod openshift-kube-controller-manager-operator_kube-controller-manager-operator-848f645654-rmdb8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada" Netns:"/var/run/netns/df186742-b74d-4a28-bd17-a8e51d863583" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager-operator;K8S_POD_NAME=kube-controller-manager-operator-848f645654-rmdb8;K8S_POD_INFRA_CONTAINER_ID=246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada;K8S_POD_UID=11f563d5-89bb-433c-956a-6d5d2492e8f1" Path:"" ERRORED: error configuring pod [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8] networking: [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8/11f563d5-89bb-433c-956a-6d5d2492e8f1:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.258861 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.258861 master-0 kubenswrapper[4752]: > pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:14.258861 master-0 kubenswrapper[4752]: E1205 10:37:14.258248 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.258861 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator_11f563d5-89bb-433c-956a-6d5d2492e8f1_0(246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada): error adding pod openshift-kube-controller-manager-operator_kube-controller-manager-operator-848f645654-rmdb8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada" Netns:"/var/run/netns/df186742-b74d-4a28-bd17-a8e51d863583" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager-operator;K8S_POD_NAME=kube-controller-manager-operator-848f645654-rmdb8;K8S_POD_INFRA_CONTAINER_ID=246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada;K8S_POD_UID=11f563d5-89bb-433c-956a-6d5d2492e8f1" Path:"" ERRORED: error configuring pod [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8] networking: [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8/11f563d5-89bb-433c-956a-6d5d2492e8f1:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.258861 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.258861 master-0 kubenswrapper[4752]: > pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:14.259113 master-0 kubenswrapper[4752]: E1205 10:37:14.258300 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator(11f563d5-89bb-433c-956a-6d5d2492e8f1)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator(11f563d5-89bb-433c-956a-6d5d2492e8f1)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_kube-controller-manager-operator-848f645654-rmdb8_openshift-kube-controller-manager-operator_11f563d5-89bb-433c-956a-6d5d2492e8f1_0(246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada): error adding pod openshift-kube-controller-manager-operator_kube-controller-manager-operator-848f645654-rmdb8 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada\\\" Netns:\\\"/var/run/netns/df186742-b74d-4a28-bd17-a8e51d863583\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager-operator;K8S_POD_NAME=kube-controller-manager-operator-848f645654-rmdb8;K8S_POD_INFRA_CONTAINER_ID=246d13096f4f47d806b7ca1bbe78eb6a0e35420a7c23f854b7e4a33686221ada;K8S_POD_UID=11f563d5-89bb-433c-956a-6d5d2492e8f1\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8] networking: [openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8/11f563d5-89bb-433c-956a-6d5d2492e8f1:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" podUID="11f563d5-89bb-433c-956a-6d5d2492e8f1" Dec 05 10:37:14.282851 master-0 kubenswrapper[4752]: E1205 10:37:14.282802 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.282851 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator_4825316a-ea9f-4d3d-838b-fa809a6e49c7_0(6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba): error adding pod openshift-kube-scheduler-operator_openshift-kube-scheduler-operator-5f85974995-dwh5t to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba" Netns:"/var/run/netns/bb08f90c-ea21-4888-a7ad-5fef85d02a73" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-scheduler-operator;K8S_POD_NAME=openshift-kube-scheduler-operator-5f85974995-dwh5t;K8S_POD_INFRA_CONTAINER_ID=6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba;K8S_POD_UID=4825316a-ea9f-4d3d-838b-fa809a6e49c7" Path:"" ERRORED: error configuring pod [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t] networking: [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t/4825316a-ea9f-4d3d-838b-fa809a6e49c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.282851 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.282851 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.285409 master-0 kubenswrapper[4752]: E1205 10:37:14.282883 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.285409 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator_4825316a-ea9f-4d3d-838b-fa809a6e49c7_0(6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba): error adding pod openshift-kube-scheduler-operator_openshift-kube-scheduler-operator-5f85974995-dwh5t to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba" Netns:"/var/run/netns/bb08f90c-ea21-4888-a7ad-5fef85d02a73" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-scheduler-operator;K8S_POD_NAME=openshift-kube-scheduler-operator-5f85974995-dwh5t;K8S_POD_INFRA_CONTAINER_ID=6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba;K8S_POD_UID=4825316a-ea9f-4d3d-838b-fa809a6e49c7" Path:"" ERRORED: error configuring pod [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t] networking: [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t/4825316a-ea9f-4d3d-838b-fa809a6e49c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.285409 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.285409 master-0 kubenswrapper[4752]: > pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:14.285409 master-0 kubenswrapper[4752]: E1205 10:37:14.282909 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.285409 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator_4825316a-ea9f-4d3d-838b-fa809a6e49c7_0(6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba): error adding pod openshift-kube-scheduler-operator_openshift-kube-scheduler-operator-5f85974995-dwh5t to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba" Netns:"/var/run/netns/bb08f90c-ea21-4888-a7ad-5fef85d02a73" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-scheduler-operator;K8S_POD_NAME=openshift-kube-scheduler-operator-5f85974995-dwh5t;K8S_POD_INFRA_CONTAINER_ID=6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba;K8S_POD_UID=4825316a-ea9f-4d3d-838b-fa809a6e49c7" Path:"" ERRORED: error configuring pod [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t] networking: [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t/4825316a-ea9f-4d3d-838b-fa809a6e49c7:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.285409 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.285409 master-0 kubenswrapper[4752]: > pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:14.286157 master-0 kubenswrapper[4752]: E1205 10:37:14.282989 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator(4825316a-ea9f-4d3d-838b-fa809a6e49c7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator(4825316a-ea9f-4d3d-838b-fa809a6e49c7)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-kube-scheduler-operator-5f85974995-dwh5t_openshift-kube-scheduler-operator_4825316a-ea9f-4d3d-838b-fa809a6e49c7_0(6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba): error adding pod openshift-kube-scheduler-operator_openshift-kube-scheduler-operator-5f85974995-dwh5t to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba\\\" Netns:\\\"/var/run/netns/bb08f90c-ea21-4888-a7ad-5fef85d02a73\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-scheduler-operator;K8S_POD_NAME=openshift-kube-scheduler-operator-5f85974995-dwh5t;K8S_POD_INFRA_CONTAINER_ID=6e6b1407e6071c00fe86e46f8f8a3e65c86400a4bd062b3f6a6acaa210ad27ba;K8S_POD_UID=4825316a-ea9f-4d3d-838b-fa809a6e49c7\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t] networking: [openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t/4825316a-ea9f-4d3d-838b-fa809a6e49c7:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" podUID="4825316a-ea9f-4d3d-838b-fa809a6e49c7" Dec 05 10:37:14.286157 master-0 kubenswrapper[4752]: E1205 10:37:14.284890 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.286157 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator_87909f47-f2d7-46f8-a1c8-27336cdcce5d_0(1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0): error adding pod openshift-cluster-storage-operator_csi-snapshot-controller-operator-6bc8656fdc-vd94f to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0" Netns:"/var/run/netns/6c660008-8764-4c4b-a6fd-7044adacfa61" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-storage-operator;K8S_POD_NAME=csi-snapshot-controller-operator-6bc8656fdc-vd94f;K8S_POD_INFRA_CONTAINER_ID=1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0;K8S_POD_UID=87909f47-f2d7-46f8-a1c8-27336cdcce5d" Path:"" ERRORED: error configuring pod [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f] networking: [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f/87909f47-f2d7-46f8-a1c8-27336cdcce5d:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.286157 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.286157 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.286157 master-0 kubenswrapper[4752]: E1205 10:37:14.284953 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.286157 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator_87909f47-f2d7-46f8-a1c8-27336cdcce5d_0(1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0): error adding pod openshift-cluster-storage-operator_csi-snapshot-controller-operator-6bc8656fdc-vd94f to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0" Netns:"/var/run/netns/6c660008-8764-4c4b-a6fd-7044adacfa61" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-storage-operator;K8S_POD_NAME=csi-snapshot-controller-operator-6bc8656fdc-vd94f;K8S_POD_INFRA_CONTAINER_ID=1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0;K8S_POD_UID=87909f47-f2d7-46f8-a1c8-27336cdcce5d" Path:"" ERRORED: error configuring pod [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f] networking: [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f/87909f47-f2d7-46f8-a1c8-27336cdcce5d:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.286635 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.286635 master-0 kubenswrapper[4752]: > pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:14.286635 master-0 kubenswrapper[4752]: E1205 10:37:14.284971 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.286635 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator_87909f47-f2d7-46f8-a1c8-27336cdcce5d_0(1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0): error adding pod openshift-cluster-storage-operator_csi-snapshot-controller-operator-6bc8656fdc-vd94f to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0" Netns:"/var/run/netns/6c660008-8764-4c4b-a6fd-7044adacfa61" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-storage-operator;K8S_POD_NAME=csi-snapshot-controller-operator-6bc8656fdc-vd94f;K8S_POD_INFRA_CONTAINER_ID=1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0;K8S_POD_UID=87909f47-f2d7-46f8-a1c8-27336cdcce5d" Path:"" ERRORED: error configuring pod [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f] networking: [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f/87909f47-f2d7-46f8-a1c8-27336cdcce5d:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.286635 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.286635 master-0 kubenswrapper[4752]: > pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:14.286635 master-0 kubenswrapper[4752]: E1205 10:37:14.285026 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator(87909f47-f2d7-46f8-a1c8-27336cdcce5d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator(87909f47-f2d7-46f8-a1c8-27336cdcce5d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_csi-snapshot-controller-operator-6bc8656fdc-vd94f_openshift-cluster-storage-operator_87909f47-f2d7-46f8-a1c8-27336cdcce5d_0(1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0): error adding pod openshift-cluster-storage-operator_csi-snapshot-controller-operator-6bc8656fdc-vd94f to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0\\\" Netns:\\\"/var/run/netns/6c660008-8764-4c4b-a6fd-7044adacfa61\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-storage-operator;K8S_POD_NAME=csi-snapshot-controller-operator-6bc8656fdc-vd94f;K8S_POD_INFRA_CONTAINER_ID=1ffc1edd0131237066ac7ab445e5879e721b12aa5a60da2d3ef05afda32fb6c0;K8S_POD_UID=87909f47-f2d7-46f8-a1c8-27336cdcce5d\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f] networking: [openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f/87909f47-f2d7-46f8-a1c8-27336cdcce5d:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" podUID="87909f47-f2d7-46f8-a1c8-27336cdcce5d" Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: E1205 10:37:14.290415 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator_6f76d12f-5406-47e2-8337-2f50e35376d6_0(23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c): error adding pod openshift-apiserver-operator_openshift-apiserver-operator-7bf7f6b755-hdjv7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c" Netns:"/var/run/netns/adbb2042-7f5a-4189-80a5-56147d9c4196" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-apiserver-operator;K8S_POD_NAME=openshift-apiserver-operator-7bf7f6b755-hdjv7;K8S_POD_INFRA_CONTAINER_ID=23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c;K8S_POD_UID=6f76d12f-5406-47e2-8337-2f50e35376d6" Path:"" ERRORED: error configuring pod [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7] networking: [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7/6f76d12f-5406-47e2-8337-2f50e35376d6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: E1205 10:37:14.290496 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator_6f76d12f-5406-47e2-8337-2f50e35376d6_0(23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c): error adding pod openshift-apiserver-operator_openshift-apiserver-operator-7bf7f6b755-hdjv7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c" Netns:"/var/run/netns/adbb2042-7f5a-4189-80a5-56147d9c4196" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-apiserver-operator;K8S_POD_NAME=openshift-apiserver-operator-7bf7f6b755-hdjv7;K8S_POD_INFRA_CONTAINER_ID=23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c;K8S_POD_UID=6f76d12f-5406-47e2-8337-2f50e35376d6" Path:"" ERRORED: error configuring pod [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7] networking: [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7/6f76d12f-5406-47e2-8337-2f50e35376d6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: > pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: E1205 10:37:14.290520 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator_6f76d12f-5406-47e2-8337-2f50e35376d6_0(23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c): error adding pod openshift-apiserver-operator_openshift-apiserver-operator-7bf7f6b755-hdjv7 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c" Netns:"/var/run/netns/adbb2042-7f5a-4189-80a5-56147d9c4196" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-apiserver-operator;K8S_POD_NAME=openshift-apiserver-operator-7bf7f6b755-hdjv7;K8S_POD_INFRA_CONTAINER_ID=23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c;K8S_POD_UID=6f76d12f-5406-47e2-8337-2f50e35376d6" Path:"" ERRORED: error configuring pod [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7] networking: [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7/6f76d12f-5406-47e2-8337-2f50e35376d6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.290538 master-0 kubenswrapper[4752]: > pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:14.291201 master-0 kubenswrapper[4752]: E1205 10:37:14.290618 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator(6f76d12f-5406-47e2-8337-2f50e35376d6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator(6f76d12f-5406-47e2-8337-2f50e35376d6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-apiserver-operator-7bf7f6b755-hdjv7_openshift-apiserver-operator_6f76d12f-5406-47e2-8337-2f50e35376d6_0(23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c): error adding pod openshift-apiserver-operator_openshift-apiserver-operator-7bf7f6b755-hdjv7 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c\\\" Netns:\\\"/var/run/netns/adbb2042-7f5a-4189-80a5-56147d9c4196\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-apiserver-operator;K8S_POD_NAME=openshift-apiserver-operator-7bf7f6b755-hdjv7;K8S_POD_INFRA_CONTAINER_ID=23a473da2dce1515b2a06d3b42516c4a6eb5d0c47812c1f3ff01059c2135c99c;K8S_POD_UID=6f76d12f-5406-47e2-8337-2f50e35376d6\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7] networking: [openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7/6f76d12f-5406-47e2-8337-2f50e35376d6:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" podUID="6f76d12f-5406-47e2-8337-2f50e35376d6" Dec 05 10:37:14.293986 master-0 kubenswrapper[4752]: E1205 10:37:14.293938 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.293986 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator_49051e6e-5a2f-45c8-bad0-374514a91c07_0(3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79): error adding pod openshift-cluster-olm-operator_cluster-olm-operator-56fcb6cc5f-m6p27 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79" Netns:"/var/run/netns/e0ef96a2-a5a1-468f-bee9-9a70e14f8786" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-olm-operator;K8S_POD_NAME=cluster-olm-operator-56fcb6cc5f-m6p27;K8S_POD_INFRA_CONTAINER_ID=3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79;K8S_POD_UID=49051e6e-5a2f-45c8-bad0-374514a91c07" Path:"" ERRORED: error configuring pod [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27] networking: [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27/49051e6e-5a2f-45c8-bad0-374514a91c07:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.293986 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.293986 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.294146 master-0 kubenswrapper[4752]: E1205 10:37:14.294006 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.294146 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator_49051e6e-5a2f-45c8-bad0-374514a91c07_0(3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79): error adding pod openshift-cluster-olm-operator_cluster-olm-operator-56fcb6cc5f-m6p27 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79" Netns:"/var/run/netns/e0ef96a2-a5a1-468f-bee9-9a70e14f8786" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-olm-operator;K8S_POD_NAME=cluster-olm-operator-56fcb6cc5f-m6p27;K8S_POD_INFRA_CONTAINER_ID=3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79;K8S_POD_UID=49051e6e-5a2f-45c8-bad0-374514a91c07" Path:"" ERRORED: error configuring pod [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27] networking: [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27/49051e6e-5a2f-45c8-bad0-374514a91c07:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.294146 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.294146 master-0 kubenswrapper[4752]: > pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:14.294146 master-0 kubenswrapper[4752]: E1205 10:37:14.294023 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.294146 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator_49051e6e-5a2f-45c8-bad0-374514a91c07_0(3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79): error adding pod openshift-cluster-olm-operator_cluster-olm-operator-56fcb6cc5f-m6p27 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79" Netns:"/var/run/netns/e0ef96a2-a5a1-468f-bee9-9a70e14f8786" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-olm-operator;K8S_POD_NAME=cluster-olm-operator-56fcb6cc5f-m6p27;K8S_POD_INFRA_CONTAINER_ID=3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79;K8S_POD_UID=49051e6e-5a2f-45c8-bad0-374514a91c07" Path:"" ERRORED: error configuring pod [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27] networking: [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27/49051e6e-5a2f-45c8-bad0-374514a91c07:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.294146 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.294146 master-0 kubenswrapper[4752]: > pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:14.294356 master-0 kubenswrapper[4752]: E1205 10:37:14.294081 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator(49051e6e-5a2f-45c8-bad0-374514a91c07)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator(49051e6e-5a2f-45c8-bad0-374514a91c07)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_cluster-olm-operator-56fcb6cc5f-m6p27_openshift-cluster-olm-operator_49051e6e-5a2f-45c8-bad0-374514a91c07_0(3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79): error adding pod openshift-cluster-olm-operator_cluster-olm-operator-56fcb6cc5f-m6p27 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79\\\" Netns:\\\"/var/run/netns/e0ef96a2-a5a1-468f-bee9-9a70e14f8786\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-cluster-olm-operator;K8S_POD_NAME=cluster-olm-operator-56fcb6cc5f-m6p27;K8S_POD_INFRA_CONTAINER_ID=3420ec623541db3f5834d76b8a75ab49dff45f69078c94c959dae0edf9d56b79;K8S_POD_UID=49051e6e-5a2f-45c8-bad0-374514a91c07\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27] networking: [openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27/49051e6e-5a2f-45c8-bad0-374514a91c07:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" podUID="49051e6e-5a2f-45c8-bad0-374514a91c07" Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: E1205 10:37:14.306864 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator_1e69ce9e-4e6f-4015-9ba6-5a7942570190_0(8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e): error adding pod openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e" Netns:"/var/run/netns/028909c0-3e6d-497f-be52-f4f7af45a064" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager-operator;K8S_POD_NAME=openshift-controller-manager-operator-6c8676f99d-cwvk5;K8S_POD_INFRA_CONTAINER_ID=8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e;K8S_POD_UID=1e69ce9e-4e6f-4015-9ba6-5a7942570190" Path:"" ERRORED: error configuring pod [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5] networking: [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5/1e69ce9e-4e6f-4015-9ba6-5a7942570190:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: E1205 10:37:14.306922 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator_1e69ce9e-4e6f-4015-9ba6-5a7942570190_0(8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e): error adding pod openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e" Netns:"/var/run/netns/028909c0-3e6d-497f-be52-f4f7af45a064" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager-operator;K8S_POD_NAME=openshift-controller-manager-operator-6c8676f99d-cwvk5;K8S_POD_INFRA_CONTAINER_ID=8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e;K8S_POD_UID=1e69ce9e-4e6f-4015-9ba6-5a7942570190" Path:"" ERRORED: error configuring pod [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5] networking: [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5/1e69ce9e-4e6f-4015-9ba6-5a7942570190:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: > pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: E1205 10:37:14.306947 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator_1e69ce9e-4e6f-4015-9ba6-5a7942570190_0(8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e): error adding pod openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e" Netns:"/var/run/netns/028909c0-3e6d-497f-be52-f4f7af45a064" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager-operator;K8S_POD_NAME=openshift-controller-manager-operator-6c8676f99d-cwvk5;K8S_POD_INFRA_CONTAINER_ID=8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e;K8S_POD_UID=1e69ce9e-4e6f-4015-9ba6-5a7942570190" Path:"" ERRORED: error configuring pod [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5] networking: [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5/1e69ce9e-4e6f-4015-9ba6-5a7942570190:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.307154 master-0 kubenswrapper[4752]: > pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:14.307628 master-0 kubenswrapper[4752]: E1205 10:37:14.307003 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator(1e69ce9e-4e6f-4015-9ba6-5a7942570190)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator(1e69ce9e-4e6f-4015-9ba6-5a7942570190)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator_1e69ce9e-4e6f-4015-9ba6-5a7942570190_0(8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e): error adding pod openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e\\\" Netns:\\\"/var/run/netns/028909c0-3e6d-497f-be52-f4f7af45a064\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager-operator;K8S_POD_NAME=openshift-controller-manager-operator-6c8676f99d-cwvk5;K8S_POD_INFRA_CONTAINER_ID=8750fe50690d3c47ea04f85182f3fd2b149bb27ddc329c3bcf66a7c4a910b45e;K8S_POD_UID=1e69ce9e-4e6f-4015-9ba6-5a7942570190\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5] networking: [openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5/1e69ce9e-4e6f-4015-9ba6-5a7942570190:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" podUID="1e69ce9e-4e6f-4015-9ba6-5a7942570190" Dec 05 10:37:14.309348 master-0 kubenswrapper[4752]: E1205 10:37:14.309313 4752 log.go:32] "RunPodSandbox from runtime service failed" err=< Dec 05 10:37:14.309348 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator_d95a56ba-c940-4e3e-aed6-d8c04f1871b6_0(bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a): error adding pod openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a" Netns:"/var/run/netns/8fbbd7df-0eb5-4fdc-941d-7c312aebd746" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication-operator;K8S_POD_NAME=authentication-operator-6c968fdfdf-t7sl8;K8S_POD_INFRA_CONTAINER_ID=bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a;K8S_POD_UID=d95a56ba-c940-4e3e-aed6-d8c04f1871b6" Path:"" ERRORED: error configuring pod [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8] networking: [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8/d95a56ba-c940-4e3e-aed6-d8c04f1871b6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.309348 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.309348 master-0 kubenswrapper[4752]: > Dec 05 10:37:14.309511 master-0 kubenswrapper[4752]: E1205 10:37:14.309354 4752 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Dec 05 10:37:14.309511 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator_d95a56ba-c940-4e3e-aed6-d8c04f1871b6_0(bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a): error adding pod openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a" Netns:"/var/run/netns/8fbbd7df-0eb5-4fdc-941d-7c312aebd746" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication-operator;K8S_POD_NAME=authentication-operator-6c968fdfdf-t7sl8;K8S_POD_INFRA_CONTAINER_ID=bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a;K8S_POD_UID=d95a56ba-c940-4e3e-aed6-d8c04f1871b6" Path:"" ERRORED: error configuring pod [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8] networking: [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8/d95a56ba-c940-4e3e-aed6-d8c04f1871b6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.309511 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.309511 master-0 kubenswrapper[4752]: > pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:14.309511 master-0 kubenswrapper[4752]: E1205 10:37:14.309369 4752 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Dec 05 10:37:14.309511 master-0 kubenswrapper[4752]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator_d95a56ba-c940-4e3e-aed6-d8c04f1871b6_0(bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a): error adding pod openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8 to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a" Netns:"/var/run/netns/8fbbd7df-0eb5-4fdc-941d-7c312aebd746" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication-operator;K8S_POD_NAME=authentication-operator-6c968fdfdf-t7sl8;K8S_POD_INFRA_CONTAINER_ID=bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a;K8S_POD_UID=d95a56ba-c940-4e3e-aed6-d8c04f1871b6" Path:"" ERRORED: error configuring pod [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8] networking: [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8/d95a56ba-c940-4e3e-aed6-d8c04f1871b6:ovn-kubernetes]: error adding container to network "ovn-kubernetes": failed to send CNI request: Post "http://dummy/": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused Dec 05 10:37:14.309511 master-0 kubenswrapper[4752]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Dec 05 10:37:14.309511 master-0 kubenswrapper[4752]: > pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:14.309716 master-0 kubenswrapper[4752]: E1205 10:37:14.309438 4752 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator(d95a56ba-c940-4e3e-aed6-d8c04f1871b6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator(d95a56ba-c940-4e3e-aed6-d8c04f1871b6)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_authentication-operator-6c968fdfdf-t7sl8_openshift-authentication-operator_d95a56ba-c940-4e3e-aed6-d8c04f1871b6_0(bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a): error adding pod openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8 to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a\\\" Netns:\\\"/var/run/netns/8fbbd7df-0eb5-4fdc-941d-7c312aebd746\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-authentication-operator;K8S_POD_NAME=authentication-operator-6c968fdfdf-t7sl8;K8S_POD_INFRA_CONTAINER_ID=bfa791178c4de4020e3cf5258b4a8e921e9dfc657a67331584a783a9681ab59a;K8S_POD_UID=d95a56ba-c940-4e3e-aed6-d8c04f1871b6\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8] networking: [openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8/d95a56ba-c940-4e3e-aed6-d8c04f1871b6:ovn-kubernetes]: error adding container to network \\\"ovn-kubernetes\\\": failed to send CNI request: Post \\\"http://dummy/\\\": dial unix /var/run/ovn-kubernetes/cni//ovn-cni-server.sock: connect: connection refused\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" podUID="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" Dec 05 10:37:15.268822 master-0 kubenswrapper[4752]: I1205 10:37:15.268748 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-d6wjk" event={"ID":"283122ba-be1c-4516-bd0f-df41c13c098b","Type":"ContainerStarted","Data":"83a3ac5e727695fe740fccfb2b719fc3eaa377e0c9d2da38823ccc4530a5c4e4"} Dec 05 10:37:15.296006 master-0 kubenswrapper[4752]: I1205 10:37:15.295931 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-operator/iptables-alerter-d6wjk" podStartSLOduration=4.346934924 podStartE2EDuration="7.295907681s" podCreationTimestamp="2025-12-05 10:37:08 +0000 UTC" firstStartedPulling="2025-12-05 10:37:08.834807935 +0000 UTC m=+110.374931264" lastFinishedPulling="2025-12-05 10:37:11.783780672 +0000 UTC m=+113.323904021" observedRunningTime="2025-12-05 10:37:15.295199594 +0000 UTC m=+116.835322983" watchObservedRunningTime="2025-12-05 10:37:15.295907681 +0000 UTC m=+116.836031020" Dec 05 10:37:16.233867 master-0 kubenswrapper[4752]: I1205 10:37:16.233646 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:16.233867 master-0 kubenswrapper[4752]: I1205 10:37:16.233750 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:16.233867 master-0 kubenswrapper[4752]: E1205 10:37:16.233894 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/node-tuning-operator-tls: secret "node-tuning-operator-tls" not found Dec 05 10:37:16.234412 master-0 kubenswrapper[4752]: E1205 10:37:16.233978 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:24.233957478 +0000 UTC m=+125.774080807 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "node-tuning-operator-tls" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "node-tuning-operator-tls" not found Dec 05 10:37:16.234412 master-0 kubenswrapper[4752]: I1205 10:37:16.234403 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:16.234697 master-0 kubenswrapper[4752]: I1205 10:37:16.234504 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:16.234697 master-0 kubenswrapper[4752]: I1205 10:37:16.234545 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:16.234697 master-0 kubenswrapper[4752]: I1205 10:37:16.234572 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:16.234697 master-0 kubenswrapper[4752]: I1205 10:37:16.234610 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:16.234697 master-0 kubenswrapper[4752]: I1205 10:37:16.234643 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:16.234697 master-0 kubenswrapper[4752]: I1205 10:37:16.234670 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:16.234697 master-0 kubenswrapper[4752]: I1205 10:37:16.234701 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:16.235246 master-0 kubenswrapper[4752]: E1205 10:37:16.234793 4752 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: secret "multus-admission-controller-secret" not found Dec 05 10:37:16.235246 master-0 kubenswrapper[4752]: E1205 10:37:16.234833 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs podName:8d76404b-6d62-4a61-b6f6-0c8073eba198 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:24.234819111 +0000 UTC m=+125.774942440 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs") pod "multus-admission-controller-7dfc5b745f-67rx7" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198") : secret "multus-admission-controller-secret" not found Dec 05 10:37:16.235246 master-0 kubenswrapper[4752]: E1205 10:37:16.235067 4752 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:16.235246 master-0 kubenswrapper[4752]: E1205 10:37:16.235130 4752 secret.go:189] Couldn't get secret openshift-image-registry/image-registry-operator-tls: secret "image-registry-operator-tls" not found Dec 05 10:37:16.235246 master-0 kubenswrapper[4752]: E1205 10:37:16.235204 4752 secret.go:189] Couldn't get secret openshift-ingress-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:16.235246 master-0 kubenswrapper[4752]: E1205 10:37:16.235210 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls podName:a722cda9-29a0-4b7f-8e1d-9a8950ed765a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:24.23517323 +0000 UTC m=+125.775296599 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-7ff994598c-kq8qr" (UID: "a722cda9-29a0-4b7f-8e1d-9a8950ed765a") : secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:16.235748 master-0 kubenswrapper[4752]: E1205 10:37:16.235271 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls podName:f7b29f89-e42d-4e53-ad14-05efdce933f0 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:24.235254092 +0000 UTC m=+125.775377501 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls") pod "cluster-image-registry-operator-6fb9f88b7-f29mb" (UID: "f7b29f89-e42d-4e53-ad14-05efdce933f0") : secret "image-registry-operator-tls" not found Dec 05 10:37:16.235748 master-0 kubenswrapper[4752]: E1205 10:37:16.235305 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls podName:22676fac-b770-4937-9bee-7478bd1babb7 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:24.235287542 +0000 UTC m=+125.775410931 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls") pod "ingress-operator-8649c48786-cgt5x" (UID: "22676fac-b770-4937-9bee-7478bd1babb7") : secret "metrics-tls" not found Dec 05 10:37:16.235748 master-0 kubenswrapper[4752]: E1205 10:37:16.235351 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: secret "package-server-manager-serving-cert" not found Dec 05 10:37:16.235748 master-0 kubenswrapper[4752]: E1205 10:37:16.235474 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert podName:8c649a16-c187-412e-b5da-62a00bee38ab nodeName:}" failed. No retries permitted until 2025-12-05 10:37:24.235410255 +0000 UTC m=+125.775533624 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert") pod "package-server-manager-67477646d4-nm8cn" (UID: "8c649a16-c187-412e-b5da-62a00bee38ab") : secret "package-server-manager-serving-cert" not found Dec 05 10:37:16.235748 master-0 kubenswrapper[4752]: E1205 10:37:16.235552 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: secret "olm-operator-serving-cert" not found Dec 05 10:37:16.235748 master-0 kubenswrapper[4752]: E1205 10:37:16.235630 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert podName:46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:24.23560197 +0000 UTC m=+125.775725339 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert") pod "olm-operator-7cd7dbb44c-d25sk" (UID: "46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082") : secret "olm-operator-serving-cert" not found Dec 05 10:37:16.235748 master-0 kubenswrapper[4752]: E1205 10:37:16.235656 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/performance-addon-operator-webhook-cert: secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:16.235748 master-0 kubenswrapper[4752]: E1205 10:37:16.235729 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:24.235709043 +0000 UTC m=+125.775832402 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:16.236513 master-0 kubenswrapper[4752]: E1205 10:37:16.235825 4752 secret.go:189] Couldn't get secret openshift-dns-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:16.236513 master-0 kubenswrapper[4752]: E1205 10:37:16.235898 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls podName:5a1bdc70-6412-47e0-8330-04d796cc8d55 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:24.235877068 +0000 UTC m=+125.776000437 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls") pod "dns-operator-7c56cf9b74-x6t9h" (UID: "5a1bdc70-6412-47e0-8330-04d796cc8d55") : secret "metrics-tls" not found Dec 05 10:37:16.236513 master-0 kubenswrapper[4752]: E1205 10:37:16.235950 4752 secret.go:189] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: secret "marketplace-operator-metrics" not found Dec 05 10:37:16.236513 master-0 kubenswrapper[4752]: E1205 10:37:16.236041 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics podName:926263c4-ec5b-41cb-9c30-0c88f636035f nodeName:}" failed. No retries permitted until 2025-12-05 10:37:24.236021942 +0000 UTC m=+125.776145331 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics") pod "marketplace-operator-f797b99b6-z9qcl" (UID: "926263c4-ec5b-41cb-9c30-0c88f636035f") : secret "marketplace-operator-metrics" not found Dec 05 10:37:24.122194 master-0 kubenswrapper[4752]: I1205 10:37:24.121841 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:24.124715 master-0 kubenswrapper[4752]: I1205 10:37:24.124632 4752 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 10:37:24.133003 master-0 kubenswrapper[4752]: E1205 10:37:24.132940 4752 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Dec 05 10:37:24.133145 master-0 kubenswrapper[4752]: E1205 10:37:24.133042 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:38:28.133009424 +0000 UTC m=+189.673132783 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : secret "metrics-daemon-secret" not found Dec 05 10:37:24.325240 master-0 kubenswrapper[4752]: I1205 10:37:24.325172 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:24.325240 master-0 kubenswrapper[4752]: I1205 10:37:24.325242 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:24.325553 master-0 kubenswrapper[4752]: I1205 10:37:24.325276 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:24.325553 master-0 kubenswrapper[4752]: I1205 10:37:24.325302 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:24.325553 master-0 kubenswrapper[4752]: I1205 10:37:24.325325 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:24.325553 master-0 kubenswrapper[4752]: I1205 10:37:24.325356 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:24.325553 master-0 kubenswrapper[4752]: E1205 10:37:24.325377 4752 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:24.325553 master-0 kubenswrapper[4752]: E1205 10:37:24.325471 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls podName:a722cda9-29a0-4b7f-8e1d-9a8950ed765a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:40.325451029 +0000 UTC m=+141.865574358 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-7ff994598c-kq8qr" (UID: "a722cda9-29a0-4b7f-8e1d-9a8950ed765a") : secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:24.325553 master-0 kubenswrapper[4752]: E1205 10:37:24.325484 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: secret "package-server-manager-serving-cert" not found Dec 05 10:37:24.325553 master-0 kubenswrapper[4752]: E1205 10:37:24.325528 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert podName:8c649a16-c187-412e-b5da-62a00bee38ab nodeName:}" failed. No retries permitted until 2025-12-05 10:37:40.325518461 +0000 UTC m=+141.865641780 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert") pod "package-server-manager-67477646d4-nm8cn" (UID: "8c649a16-c187-412e-b5da-62a00bee38ab") : secret "package-server-manager-serving-cert" not found Dec 05 10:37:24.325846 master-0 kubenswrapper[4752]: E1205 10:37:24.325555 4752 secret.go:189] Couldn't get secret openshift-image-registry/image-registry-operator-tls: secret "image-registry-operator-tls" not found Dec 05 10:37:24.325846 master-0 kubenswrapper[4752]: E1205 10:37:24.325647 4752 secret.go:189] Couldn't get secret openshift-dns-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:24.325846 master-0 kubenswrapper[4752]: E1205 10:37:24.325662 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls podName:f7b29f89-e42d-4e53-ad14-05efdce933f0 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:40.325636144 +0000 UTC m=+141.865759483 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls") pod "cluster-image-registry-operator-6fb9f88b7-f29mb" (UID: "f7b29f89-e42d-4e53-ad14-05efdce933f0") : secret "image-registry-operator-tls" not found Dec 05 10:37:24.325846 master-0 kubenswrapper[4752]: I1205 10:37:24.325765 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:24.325846 master-0 kubenswrapper[4752]: E1205 10:37:24.325794 4752 secret.go:189] Couldn't get secret openshift-ingress-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:24.326053 master-0 kubenswrapper[4752]: E1205 10:37:24.325854 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/performance-addon-operator-webhook-cert: secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:24.326053 master-0 kubenswrapper[4752]: I1205 10:37:24.325809 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:24.326053 master-0 kubenswrapper[4752]: E1205 10:37:24.325874 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls podName:22676fac-b770-4937-9bee-7478bd1babb7 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:40.325849299 +0000 UTC m=+141.865972668 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls") pod "ingress-operator-8649c48786-cgt5x" (UID: "22676fac-b770-4937-9bee-7478bd1babb7") : secret "metrics-tls" not found Dec 05 10:37:24.326053 master-0 kubenswrapper[4752]: E1205 10:37:24.325902 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:40.32588967 +0000 UTC m=+141.866013039 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:24.326053 master-0 kubenswrapper[4752]: E1205 10:37:24.325911 4752 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: secret "olm-operator-serving-cert" not found Dec 05 10:37:24.326053 master-0 kubenswrapper[4752]: E1205 10:37:24.325930 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls podName:5a1bdc70-6412-47e0-8330-04d796cc8d55 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:40.325915182 +0000 UTC m=+141.866038551 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls") pod "dns-operator-7c56cf9b74-x6t9h" (UID: "5a1bdc70-6412-47e0-8330-04d796cc8d55") : secret "metrics-tls" not found Dec 05 10:37:24.326053 master-0 kubenswrapper[4752]: E1205 10:37:24.325913 4752 secret.go:189] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: secret "marketplace-operator-metrics" not found Dec 05 10:37:24.326053 master-0 kubenswrapper[4752]: E1205 10:37:24.326000 4752 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: secret "multus-admission-controller-secret" not found Dec 05 10:37:24.326053 master-0 kubenswrapper[4752]: E1205 10:37:24.325999 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert podName:46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:40.325987574 +0000 UTC m=+141.866110953 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert") pod "olm-operator-7cd7dbb44c-d25sk" (UID: "46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082") : secret "olm-operator-serving-cert" not found Dec 05 10:37:24.326053 master-0 kubenswrapper[4752]: E1205 10:37:24.326037 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs podName:8d76404b-6d62-4a61-b6f6-0c8073eba198 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:40.326029445 +0000 UTC m=+141.866152774 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs") pod "multus-admission-controller-7dfc5b745f-67rx7" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198") : secret "multus-admission-controller-secret" not found Dec 05 10:37:24.326053 master-0 kubenswrapper[4752]: I1205 10:37:24.325974 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:24.326514 master-0 kubenswrapper[4752]: E1205 10:37:24.326097 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics podName:926263c4-ec5b-41cb-9c30-0c88f636035f nodeName:}" failed. No retries permitted until 2025-12-05 10:37:40.326063286 +0000 UTC m=+141.866186705 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics") pod "marketplace-operator-f797b99b6-z9qcl" (UID: "926263c4-ec5b-41cb-9c30-0c88f636035f") : secret "marketplace-operator-metrics" not found Dec 05 10:37:24.326514 master-0 kubenswrapper[4752]: I1205 10:37:24.326156 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:24.326514 master-0 kubenswrapper[4752]: E1205 10:37:24.326307 4752 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/node-tuning-operator-tls: secret "node-tuning-operator-tls" not found Dec 05 10:37:24.326514 master-0 kubenswrapper[4752]: E1205 10:37:24.326344 4752 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:40.326336722 +0000 UTC m=+141.866460051 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "node-tuning-operator-tls" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "node-tuning-operator-tls" not found Dec 05 10:37:24.795770 master-0 kubenswrapper[4752]: I1205 10:37:24.795694 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:24.796024 master-0 kubenswrapper[4752]: I1205 10:37:24.795761 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:24.796134 master-0 kubenswrapper[4752]: I1205 10:37:24.796101 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:24.797028 master-0 kubenswrapper[4752]: I1205 10:37:24.796975 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:25.796574 master-0 kubenswrapper[4752]: I1205 10:37:25.796056 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:25.797401 master-0 kubenswrapper[4752]: I1205 10:37:25.796128 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:25.797401 master-0 kubenswrapper[4752]: I1205 10:37:25.796149 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:25.797401 master-0 kubenswrapper[4752]: I1205 10:37:25.796178 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:25.797401 master-0 kubenswrapper[4752]: I1205 10:37:25.796265 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:25.797401 master-0 kubenswrapper[4752]: I1205 10:37:25.797348 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:25.797401 master-0 kubenswrapper[4752]: I1205 10:37:25.797375 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:25.797762 master-0 kubenswrapper[4752]: I1205 10:37:25.797354 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:25.797762 master-0 kubenswrapper[4752]: I1205 10:37:25.797695 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:25.798402 master-0 kubenswrapper[4752]: I1205 10:37:25.797921 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:26.795621 master-0 kubenswrapper[4752]: I1205 10:37:26.795486 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:26.795621 master-0 kubenswrapper[4752]: I1205 10:37:26.795588 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:26.795982 master-0 kubenswrapper[4752]: I1205 10:37:26.795588 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:26.796124 master-0 kubenswrapper[4752]: I1205 10:37:26.796081 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:26.796226 master-0 kubenswrapper[4752]: I1205 10:37:26.796084 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:26.796324 master-0 kubenswrapper[4752]: I1205 10:37:26.796086 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:26.908691 master-0 kubenswrapper[4752]: I1205 10:37:26.908485 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c"] Dec 05 10:37:26.909851 master-0 kubenswrapper[4752]: I1205 10:37:26.909799 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5"] Dec 05 10:37:26.925593 master-0 kubenswrapper[4752]: W1205 10:37:26.925477 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7a08359_0379_4364_8b0c_ddb58ff605f4.slice/crio-0d2211dcab91af9edc388c398c5fc7a90c2e8c65270b4ebc9dead574b96eaefe WatchSource:0}: Error finding container 0d2211dcab91af9edc388c398c5fc7a90c2e8c65270b4ebc9dead574b96eaefe: Status 404 returned error can't find the container with id 0d2211dcab91af9edc388c398c5fc7a90c2e8c65270b4ebc9dead574b96eaefe Dec 05 10:37:27.309049 master-0 kubenswrapper[4752]: I1205 10:37:27.308997 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" event={"ID":"1e69ce9e-4e6f-4015-9ba6-5a7942570190","Type":"ContainerStarted","Data":"e2506cc6607d77e49135fcb375a37bafe2d511f6ed75718677873e40b092ab87"} Dec 05 10:37:27.310801 master-0 kubenswrapper[4752]: I1205 10:37:27.310179 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" event={"ID":"f7a08359-0379-4364-8b0c-ddb58ff605f4","Type":"ContainerStarted","Data":"0d2211dcab91af9edc388c398c5fc7a90c2e8c65270b4ebc9dead574b96eaefe"} Dec 05 10:37:27.541299 master-0 kubenswrapper[4752]: I1205 10:37:27.541258 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8"] Dec 05 10:37:27.541365 master-0 kubenswrapper[4752]: I1205 10:37:27.541319 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr"] Dec 05 10:37:27.543486 master-0 kubenswrapper[4752]: I1205 10:37:27.541629 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv"] Dec 05 10:37:27.548330 master-0 kubenswrapper[4752]: I1205 10:37:27.546036 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl"] Dec 05 10:37:27.548330 master-0 kubenswrapper[4752]: I1205 10:37:27.546809 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4"] Dec 05 10:37:28.317189 master-0 kubenswrapper[4752]: I1205 10:37:28.317127 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" event={"ID":"c22d947f-a5b6-4f24-b142-dd201c46293b","Type":"ContainerStarted","Data":"7e22b55ec9c49898cca9f69ea15552e8a230cbb6781fbf80f0e4919b22b7b7d8"} Dec 05 10:37:28.318400 master-0 kubenswrapper[4752]: I1205 10:37:28.318331 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" event={"ID":"fd58232c-a81a-4aee-8b2c-5ffcdded2e23","Type":"ContainerStarted","Data":"ccbec542fddb9bd0831349589c4ac65650d79b1cf75a0af75cf8ca2d1a0d89f6"} Dec 05 10:37:28.319086 master-0 kubenswrapper[4752]: I1205 10:37:28.319036 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" event={"ID":"d95a56ba-c940-4e3e-aed6-d8c04f1871b6","Type":"ContainerStarted","Data":"b7452830de99c0487a8d247b23da92391c3a1113eb379e2b40ac1673b8cb1fe1"} Dec 05 10:37:28.319718 master-0 kubenswrapper[4752]: I1205 10:37:28.319680 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" event={"ID":"d1c3b7dd-f25e-4983-8a94-084f863fd5b9","Type":"ContainerStarted","Data":"bdfbd9daf37e036f9b2d20314905a9aa48f68e21954c4e133ee980fe10b0d8d2"} Dec 05 10:37:28.320946 master-0 kubenswrapper[4752]: I1205 10:37:28.320886 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" event={"ID":"444f8808-e454-4015-9e20-429e715a08c7","Type":"ContainerStarted","Data":"a4e599b9d9dac0eea61bdc89656e6a7ee18f83aa3fd9ba9340f57da02bd4a39b"} Dec 05 10:37:28.320946 master-0 kubenswrapper[4752]: I1205 10:37:28.320911 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" event={"ID":"444f8808-e454-4015-9e20-429e715a08c7","Type":"ContainerStarted","Data":"8114256db0301425275a52e5c6f4316fb7a7d5eed5146413b166f9ab3918f143"} Dec 05 10:37:28.795875 master-0 kubenswrapper[4752]: I1205 10:37:28.795800 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:28.805909 master-0 kubenswrapper[4752]: I1205 10:37:28.805765 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:29.012720 master-0 kubenswrapper[4752]: I1205 10:37:29.012627 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f"] Dec 05 10:37:29.015689 master-0 kubenswrapper[4752]: I1205 10:37:29.015030 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7"] Dec 05 10:37:29.015689 master-0 kubenswrapper[4752]: I1205 10:37:29.015085 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8"] Dec 05 10:37:29.325555 master-0 kubenswrapper[4752]: I1205 10:37:29.325499 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" event={"ID":"6f76d12f-5406-47e2-8337-2f50e35376d6","Type":"ContainerStarted","Data":"fedfe3c8bf160c89d4d691575ab31b377473c05a2dd753786d2ef1a80ea5d562"} Dec 05 10:37:29.326589 master-0 kubenswrapper[4752]: I1205 10:37:29.326544 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" event={"ID":"11f563d5-89bb-433c-956a-6d5d2492e8f1","Type":"ContainerStarted","Data":"5b905a3a36c591b5f3704d14e85aafe47726c309b50985c9308b5ed0d4a90267"} Dec 05 10:37:29.327497 master-0 kubenswrapper[4752]: I1205 10:37:29.327462 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" event={"ID":"87909f47-f2d7-46f8-a1c8-27336cdcce5d","Type":"ContainerStarted","Data":"12d79e8d025da0e5c89728d8b35230c8c9764259d39417d2ff38f00dc9c263f6"} Dec 05 10:37:29.795562 master-0 kubenswrapper[4752]: I1205 10:37:29.795412 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:29.796271 master-0 kubenswrapper[4752]: I1205 10:37:29.796238 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:30.188930 master-0 kubenswrapper[4752]: I1205 10:37:30.188723 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27"] Dec 05 10:37:30.201922 master-0 kubenswrapper[4752]: W1205 10:37:30.201852 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49051e6e_5a2f_45c8_bad0_374514a91c07.slice/crio-4833f667ac4451a150ed477789be72c08b65162d8c7bbf29e44cd656299dbd63 WatchSource:0}: Error finding container 4833f667ac4451a150ed477789be72c08b65162d8c7bbf29e44cd656299dbd63: Status 404 returned error can't find the container with id 4833f667ac4451a150ed477789be72c08b65162d8c7bbf29e44cd656299dbd63 Dec 05 10:37:30.204316 master-0 kubenswrapper[4752]: I1205 10:37:30.204249 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t"] Dec 05 10:37:30.333150 master-0 kubenswrapper[4752]: I1205 10:37:30.333012 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" event={"ID":"4825316a-ea9f-4d3d-838b-fa809a6e49c7","Type":"ContainerStarted","Data":"5ab705ed0c0dfea5ee8845410f03f5084071123b3566b80a7ef7e6fcce0fa974"} Dec 05 10:37:30.334351 master-0 kubenswrapper[4752]: I1205 10:37:30.334262 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" event={"ID":"49051e6e-5a2f-45c8-bad0-374514a91c07","Type":"ContainerStarted","Data":"4833f667ac4451a150ed477789be72c08b65162d8c7bbf29e44cd656299dbd63"} Dec 05 10:37:30.487154 master-0 kubenswrapper[4752]: I1205 10:37:30.487029 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" podStartSLOduration=95.487005088 podStartE2EDuration="1m35.487005088s" podCreationTimestamp="2025-12-05 10:35:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:37:30.486152526 +0000 UTC m=+132.026275875" watchObservedRunningTime="2025-12-05 10:37:30.487005088 +0000 UTC m=+132.027128427" Dec 05 10:37:37.358458 master-0 kubenswrapper[4752]: I1205 10:37:37.358121 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" event={"ID":"f7a08359-0379-4364-8b0c-ddb58ff605f4","Type":"ContainerStarted","Data":"7f719d929cd8f822ace129247877e00bae5a9457d73d9da2566b48784c2ee8fc"} Dec 05 10:37:37.363934 master-0 kubenswrapper[4752]: I1205 10:37:37.363881 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" event={"ID":"11f563d5-89bb-433c-956a-6d5d2492e8f1","Type":"ContainerStarted","Data":"475c118c4e82c9f8f86c3034eb04eddc9ae813cf612425cd157e214036c29617"} Dec 05 10:37:37.365276 master-0 kubenswrapper[4752]: I1205 10:37:37.365217 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" event={"ID":"1e69ce9e-4e6f-4015-9ba6-5a7942570190","Type":"ContainerStarted","Data":"41f047cf59b8b167740e4f4b07cc948566d4077a26e6c889e254fc22e5a5bd11"} Dec 05 10:37:37.367284 master-0 kubenswrapper[4752]: I1205 10:37:37.367201 4752 generic.go:334] "Generic (PLEG): container finished" podID="c22d947f-a5b6-4f24-b142-dd201c46293b" containerID="d0d4427b1c533edfc1d254b4b4202c620d108a1b343169a0fecb4359d771c631" exitCode=0 Dec 05 10:37:37.367364 master-0 kubenswrapper[4752]: I1205 10:37:37.367298 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" event={"ID":"c22d947f-a5b6-4f24-b142-dd201c46293b","Type":"ContainerDied","Data":"d0d4427b1c533edfc1d254b4b4202c620d108a1b343169a0fecb4359d771c631"} Dec 05 10:37:37.370712 master-0 kubenswrapper[4752]: I1205 10:37:37.370664 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" event={"ID":"fd58232c-a81a-4aee-8b2c-5ffcdded2e23","Type":"ContainerStarted","Data":"3908741f62d9038e3367c49d1d01f504f11849e5b644100de90c2c7d751cc2e3"} Dec 05 10:37:37.374873 master-0 kubenswrapper[4752]: I1205 10:37:37.374818 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" event={"ID":"d95a56ba-c940-4e3e-aed6-d8c04f1871b6","Type":"ContainerStarted","Data":"dad5a560aae0ea7cbefd19607cf4263a0d5cfc63a06ebe4360483bee7047a720"} Dec 05 10:37:37.377742 master-0 kubenswrapper[4752]: I1205 10:37:37.377683 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" podStartSLOduration=95.498583337 podStartE2EDuration="1m43.377673151s" podCreationTimestamp="2025-12-05 10:35:54 +0000 UTC" firstStartedPulling="2025-12-05 10:37:26.926958173 +0000 UTC m=+128.467081502" lastFinishedPulling="2025-12-05 10:37:34.806047987 +0000 UTC m=+136.346171316" observedRunningTime="2025-12-05 10:37:37.377250769 +0000 UTC m=+138.917374118" watchObservedRunningTime="2025-12-05 10:37:37.377673151 +0000 UTC m=+138.917796490" Dec 05 10:37:37.378541 master-0 kubenswrapper[4752]: I1205 10:37:37.378384 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" event={"ID":"d1c3b7dd-f25e-4983-8a94-084f863fd5b9","Type":"ContainerStarted","Data":"e562ce0da1077d81598e22070b676b5642a7e1f7ad3053d9875573b9c4c32d50"} Dec 05 10:37:37.382576 master-0 kubenswrapper[4752]: I1205 10:37:37.381819 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" event={"ID":"87909f47-f2d7-46f8-a1c8-27336cdcce5d","Type":"ContainerStarted","Data":"a808bdb877e1408940941cac2293f71c7f3c08dcec207a34a49d454b85bfe548"} Dec 05 10:37:37.387246 master-0 kubenswrapper[4752]: I1205 10:37:37.387183 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" event={"ID":"6f76d12f-5406-47e2-8337-2f50e35376d6","Type":"ContainerStarted","Data":"cc8433167039cff15f2a8553e5790127434ae4c5286a5db0777d417011f80f8c"} Dec 05 10:37:37.388207 master-0 kubenswrapper[4752]: I1205 10:37:37.388172 4752 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:37.390257 master-0 kubenswrapper[4752]: I1205 10:37:37.389674 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" event={"ID":"4825316a-ea9f-4d3d-838b-fa809a6e49c7","Type":"ContainerStarted","Data":"f2785fc5d00b92168028cf4e3273a28c6085e8db64c579bdce5eca572512c7d9"} Dec 05 10:37:37.394600 master-0 kubenswrapper[4752]: I1205 10:37:37.391428 4752 generic.go:334] "Generic (PLEG): container finished" podID="49051e6e-5a2f-45c8-bad0-374514a91c07" containerID="be912c90bc4a342d66fa17918d807831fa70aa4a080f3751a563cdc9efce0a37" exitCode=0 Dec 05 10:37:37.394600 master-0 kubenswrapper[4752]: I1205 10:37:37.391460 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" event={"ID":"49051e6e-5a2f-45c8-bad0-374514a91c07","Type":"ContainerDied","Data":"be912c90bc4a342d66fa17918d807831fa70aa4a080f3751a563cdc9efce0a37"} Dec 05 10:37:37.396144 master-0 kubenswrapper[4752]: I1205 10:37:37.395399 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" podStartSLOduration=94.536254117 podStartE2EDuration="1m43.395378082s" podCreationTimestamp="2025-12-05 10:35:54 +0000 UTC" firstStartedPulling="2025-12-05 10:37:27.554313656 +0000 UTC m=+129.094436975" lastFinishedPulling="2025-12-05 10:37:36.413437611 +0000 UTC m=+137.953560940" observedRunningTime="2025-12-05 10:37:37.394830286 +0000 UTC m=+138.934953615" watchObservedRunningTime="2025-12-05 10:37:37.395378082 +0000 UTC m=+138.935501411" Dec 05 10:37:37.434248 master-0 kubenswrapper[4752]: I1205 10:37:37.434166 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" podStartSLOduration=95.064616438 podStartE2EDuration="1m43.434137688s" podCreationTimestamp="2025-12-05 10:35:54 +0000 UTC" firstStartedPulling="2025-12-05 10:37:26.937923773 +0000 UTC m=+128.478047112" lastFinishedPulling="2025-12-05 10:37:35.307445043 +0000 UTC m=+136.847568362" observedRunningTime="2025-12-05 10:37:37.43387442 +0000 UTC m=+138.973997749" watchObservedRunningTime="2025-12-05 10:37:37.434137688 +0000 UTC m=+138.974261027" Dec 05 10:37:37.481455 master-0 kubenswrapper[4752]: I1205 10:37:37.479694 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" podStartSLOduration=95.587620571 podStartE2EDuration="1m44.479671755s" podCreationTimestamp="2025-12-05 10:35:53 +0000 UTC" firstStartedPulling="2025-12-05 10:37:27.553842674 +0000 UTC m=+129.093966003" lastFinishedPulling="2025-12-05 10:37:36.445893858 +0000 UTC m=+137.986017187" observedRunningTime="2025-12-05 10:37:37.456308564 +0000 UTC m=+138.996431913" watchObservedRunningTime="2025-12-05 10:37:37.479671755 +0000 UTC m=+139.019795084" Dec 05 10:37:37.505958 master-0 kubenswrapper[4752]: I1205 10:37:37.505895 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" podStartSLOduration=95.086859244 podStartE2EDuration="1m42.505878526s" podCreationTimestamp="2025-12-05 10:35:55 +0000 UTC" firstStartedPulling="2025-12-05 10:37:29.026799134 +0000 UTC m=+130.566922473" lastFinishedPulling="2025-12-05 10:37:36.445818436 +0000 UTC m=+137.985941755" observedRunningTime="2025-12-05 10:37:37.480872219 +0000 UTC m=+139.020995548" watchObservedRunningTime="2025-12-05 10:37:37.505878526 +0000 UTC m=+139.046001855" Dec 05 10:37:37.512442 master-0 kubenswrapper[4752]: I1205 10:37:37.506611 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" podStartSLOduration=96.09839573 podStartE2EDuration="1m43.506606806s" podCreationTimestamp="2025-12-05 10:35:54 +0000 UTC" firstStartedPulling="2025-12-05 10:37:29.023308515 +0000 UTC m=+130.563431844" lastFinishedPulling="2025-12-05 10:37:36.431519591 +0000 UTC m=+137.971642920" observedRunningTime="2025-12-05 10:37:37.505310119 +0000 UTC m=+139.045433468" watchObservedRunningTime="2025-12-05 10:37:37.506606806 +0000 UTC m=+139.046730145" Dec 05 10:37:37.532540 master-0 kubenswrapper[4752]: I1205 10:37:37.529892 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" podStartSLOduration=96.313405748 podStartE2EDuration="1m42.529871554s" podCreationTimestamp="2025-12-05 10:35:55 +0000 UTC" firstStartedPulling="2025-12-05 10:37:30.217214147 +0000 UTC m=+131.757337506" lastFinishedPulling="2025-12-05 10:37:36.433679983 +0000 UTC m=+137.973803312" observedRunningTime="2025-12-05 10:37:37.528363361 +0000 UTC m=+139.068486690" watchObservedRunningTime="2025-12-05 10:37:37.529871554 +0000 UTC m=+139.069994883" Dec 05 10:37:37.599244 master-0 kubenswrapper[4752]: I1205 10:37:37.596739 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" podStartSLOduration=96.178315823 podStartE2EDuration="1m43.596717954s" podCreationTimestamp="2025-12-05 10:35:54 +0000 UTC" firstStartedPulling="2025-12-05 10:37:29.022146506 +0000 UTC m=+130.562269835" lastFinishedPulling="2025-12-05 10:37:36.440548637 +0000 UTC m=+137.980671966" observedRunningTime="2025-12-05 10:37:37.595587882 +0000 UTC m=+139.135711221" watchObservedRunningTime="2025-12-05 10:37:37.596717954 +0000 UTC m=+139.136841283" Dec 05 10:37:37.634001 master-0 kubenswrapper[4752]: I1205 10:37:37.630331 4752 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" podStartSLOduration=95.812761376 podStartE2EDuration="1m44.630308574s" podCreationTimestamp="2025-12-05 10:35:53 +0000 UTC" firstStartedPulling="2025-12-05 10:37:27.557834406 +0000 UTC m=+129.097957735" lastFinishedPulling="2025-12-05 10:37:36.375381604 +0000 UTC m=+137.915504933" observedRunningTime="2025-12-05 10:37:37.629607043 +0000 UTC m=+139.169730392" watchObservedRunningTime="2025-12-05 10:37:37.630308574 +0000 UTC m=+139.170431903" Dec 05 10:37:37.755473 master-0 kubenswrapper[4752]: I1205 10:37:37.755369 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v"] Dec 05 10:37:37.757087 master-0 kubenswrapper[4752]: I1205 10:37:37.755940 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" Dec 05 10:37:37.770252 master-0 kubenswrapper[4752]: I1205 10:37:37.770207 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v"] Dec 05 10:37:37.800701 master-0 kubenswrapper[4752]: I1205 10:37:37.798387 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ml74w\" (UniqueName: \"kubernetes.io/projected/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f-kube-api-access-ml74w\") pod \"csi-snapshot-controller-6b958b6f94-lgn6v\" (UID: \"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" Dec 05 10:37:37.899124 master-0 kubenswrapper[4752]: I1205 10:37:37.898957 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml74w\" (UniqueName: \"kubernetes.io/projected/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f-kube-api-access-ml74w\") pod \"csi-snapshot-controller-6b958b6f94-lgn6v\" (UID: \"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" Dec 05 10:37:37.923051 master-0 kubenswrapper[4752]: I1205 10:37:37.922992 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml74w\" (UniqueName: \"kubernetes.io/projected/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f-kube-api-access-ml74w\") pod \"csi-snapshot-controller-6b958b6f94-lgn6v\" (UID: \"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" Dec 05 10:37:38.071908 master-0 kubenswrapper[4752]: I1205 10:37:38.071840 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" Dec 05 10:37:38.264868 master-0 kubenswrapper[4752]: I1205 10:37:38.264481 4752 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd"] Dec 05 10:37:38.266572 master-0 kubenswrapper[4752]: I1205 10:37:38.266536 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd" Dec 05 10:37:38.277888 master-0 kubenswrapper[4752]: I1205 10:37:38.273618 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v"] Dec 05 10:37:38.283384 master-0 kubenswrapper[4752]: I1205 10:37:38.282922 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 10:37:38.283384 master-0 kubenswrapper[4752]: I1205 10:37:38.283137 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 10:37:38.299272 master-0 kubenswrapper[4752]: I1205 10:37:38.299207 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd"] Dec 05 10:37:38.311365 master-0 kubenswrapper[4752]: I1205 10:37:38.311315 4752 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdsh9\" (UniqueName: \"kubernetes.io/projected/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a-kube-api-access-pdsh9\") pod \"migrator-74b7b57c65-sfvzd\" (UID: \"3c4f767d-65f9-434b-8ddd-ceb0b91ab99a\") " pod="openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd" Dec 05 10:37:38.397835 master-0 kubenswrapper[4752]: I1205 10:37:38.397784 4752 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" event={"ID":"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f","Type":"ContainerStarted","Data":"eecf6d3a967ce077414ad07bec5a21f5acf63791b8f14de4d8d34c6cfe91444a"} Dec 05 10:37:38.414252 master-0 kubenswrapper[4752]: I1205 10:37:38.414134 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdsh9\" (UniqueName: \"kubernetes.io/projected/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a-kube-api-access-pdsh9\") pod \"migrator-74b7b57c65-sfvzd\" (UID: \"3c4f767d-65f9-434b-8ddd-ceb0b91ab99a\") " pod="openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd" Dec 05 10:37:38.432374 master-0 kubenswrapper[4752]: I1205 10:37:38.432328 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdsh9\" (UniqueName: \"kubernetes.io/projected/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a-kube-api-access-pdsh9\") pod \"migrator-74b7b57c65-sfvzd\" (UID: \"3c4f767d-65f9-434b-8ddd-ceb0b91ab99a\") " pod="openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd" Dec 05 10:37:38.639422 master-0 kubenswrapper[4752]: I1205 10:37:38.638854 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd" Dec 05 10:37:38.721233 master-0 kubenswrapper[4752]: I1205 10:37:38.721164 4752 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:38.724212 master-0 kubenswrapper[4752]: I1205 10:37:38.724144 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 10:37:38.733396 master-0 kubenswrapper[4752]: I1205 10:37:38.733349 4752 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 10:37:38.754759 master-0 kubenswrapper[4752]: I1205 10:37:38.753730 4752 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:38.832866 master-0 kubenswrapper[4752]: I1205 10:37:38.832776 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd"] Dec 05 10:37:38.844006 master-0 kubenswrapper[4752]: W1205 10:37:38.843949 4752 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c4f767d_65f9_434b_8ddd_ceb0b91ab99a.slice/crio-70e7f53d39e81bfad8819ec3c2244577ab77bdd0b9a5512ba3a7cb61ae8bafb4 WatchSource:0}: Error finding container 70e7f53d39e81bfad8819ec3c2244577ab77bdd0b9a5512ba3a7cb61ae8bafb4: Status 404 returned error can't find the container with id 70e7f53d39e81bfad8819ec3c2244577ab77bdd0b9a5512ba3a7cb61ae8bafb4 Dec 05 10:37:38.909209 master-0 kubenswrapper[4752]: I1205 10:37:38.909140 4752 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:39.127024 master-0 kubenswrapper[4752]: I1205 10:37:39.126678 4752 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-target-d6fzk"] Dec 05 10:37:39.315996 master-0 kubenswrapper[4752]: I1205 10:37:39.315732 4752 dynamic_cafile_content.go:175] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 10:37:39.315942 master-0 systemd[1]: Stopping Kubernetes Kubelet... Dec 05 10:37:39.350797 master-0 systemd[1]: kubelet.service: Deactivated successfully. Dec 05 10:37:39.351104 master-0 systemd[1]: Stopped Kubernetes Kubelet. Dec 05 10:37:39.352232 master-0 systemd[1]: kubelet.service: Consumed 10.576s CPU time. Dec 05 10:37:39.355528 master-0 systemd[1]: Starting Kubernetes Kubelet... Dec 05 10:37:39.452364 master-0 kubenswrapper[10444]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:37:39.452364 master-0 kubenswrapper[10444]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 05 10:37:39.452364 master-0 kubenswrapper[10444]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:37:39.452364 master-0 kubenswrapper[10444]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:37:39.452364 master-0 kubenswrapper[10444]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 05 10:37:39.452364 master-0 kubenswrapper[10444]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:37:39.453412 master-0 kubenswrapper[10444]: I1205 10:37:39.452474 10444 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 05 10:37:39.455222 master-0 kubenswrapper[10444]: W1205 10:37:39.455186 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:37:39.455222 master-0 kubenswrapper[10444]: W1205 10:37:39.455207 10444 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:37:39.455222 master-0 kubenswrapper[10444]: W1205 10:37:39.455213 10444 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:37:39.455222 master-0 kubenswrapper[10444]: W1205 10:37:39.455217 10444 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:37:39.455222 master-0 kubenswrapper[10444]: W1205 10:37:39.455223 10444 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:37:39.455222 master-0 kubenswrapper[10444]: W1205 10:37:39.455228 10444 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455232 10444 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455237 10444 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455242 10444 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455247 10444 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455251 10444 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455256 10444 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455268 10444 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455274 10444 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455278 10444 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455283 10444 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455288 10444 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455292 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455296 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455301 10444 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455306 10444 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455310 10444 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455314 10444 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455318 10444 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455323 10444 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:37:39.455481 master-0 kubenswrapper[10444]: W1205 10:37:39.455328 10444 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455334 10444 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455340 10444 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455346 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455351 10444 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455356 10444 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455361 10444 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455366 10444 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455370 10444 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455374 10444 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455379 10444 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455383 10444 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455388 10444 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455394 10444 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455402 10444 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455407 10444 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455412 10444 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455417 10444 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455426 10444 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:37:39.456193 master-0 kubenswrapper[10444]: W1205 10:37:39.455448 10444 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455454 10444 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455460 10444 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455465 10444 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455469 10444 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455474 10444 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455478 10444 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455484 10444 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455489 10444 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455493 10444 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455497 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455503 10444 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455508 10444 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455512 10444 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455517 10444 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455523 10444 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455527 10444 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455531 10444 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455536 10444 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455540 10444 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:37:39.456822 master-0 kubenswrapper[10444]: W1205 10:37:39.455545 10444 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: W1205 10:37:39.455552 10444 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: W1205 10:37:39.455557 10444 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: W1205 10:37:39.455563 10444 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: W1205 10:37:39.455568 10444 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: W1205 10:37:39.455579 10444 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: W1205 10:37:39.455583 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: W1205 10:37:39.455588 10444 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455718 10444 flags.go:64] FLAG: --address="0.0.0.0" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455732 10444 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455742 10444 flags.go:64] FLAG: --anonymous-auth="true" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455749 10444 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455757 10444 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455762 10444 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455770 10444 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455775 10444 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455781 10444 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455786 10444 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455793 10444 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455798 10444 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455803 10444 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455808 10444 flags.go:64] FLAG: --cgroup-root="" Dec 05 10:37:39.457481 master-0 kubenswrapper[10444]: I1205 10:37:39.455813 10444 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455818 10444 flags.go:64] FLAG: --client-ca-file="" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455824 10444 flags.go:64] FLAG: --cloud-config="" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455829 10444 flags.go:64] FLAG: --cloud-provider="" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455834 10444 flags.go:64] FLAG: --cluster-dns="[]" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455840 10444 flags.go:64] FLAG: --cluster-domain="" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455845 10444 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455850 10444 flags.go:64] FLAG: --config-dir="" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455854 10444 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455860 10444 flags.go:64] FLAG: --container-log-max-files="5" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455867 10444 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455872 10444 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455878 10444 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455883 10444 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455889 10444 flags.go:64] FLAG: --contention-profiling="false" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455898 10444 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455903 10444 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455908 10444 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455913 10444 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455920 10444 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455925 10444 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455929 10444 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455934 10444 flags.go:64] FLAG: --enable-load-reader="false" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455939 10444 flags.go:64] FLAG: --enable-server="true" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455944 10444 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 05 10:37:39.458256 master-0 kubenswrapper[10444]: I1205 10:37:39.455952 10444 flags.go:64] FLAG: --event-burst="100" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.455957 10444 flags.go:64] FLAG: --event-qps="50" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.455963 10444 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.455968 10444 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.455974 10444 flags.go:64] FLAG: --eviction-hard="" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.455980 10444 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.455986 10444 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.455992 10444 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.455997 10444 flags.go:64] FLAG: --eviction-soft="" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456003 10444 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456007 10444 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456013 10444 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456018 10444 flags.go:64] FLAG: --experimental-mounter-path="" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456023 10444 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456028 10444 flags.go:64] FLAG: --fail-swap-on="true" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456034 10444 flags.go:64] FLAG: --feature-gates="" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456040 10444 flags.go:64] FLAG: --file-check-frequency="20s" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456047 10444 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456052 10444 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456058 10444 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456063 10444 flags.go:64] FLAG: --healthz-port="10248" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456069 10444 flags.go:64] FLAG: --help="false" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456077 10444 flags.go:64] FLAG: --hostname-override="" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456083 10444 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456088 10444 flags.go:64] FLAG: --http-check-frequency="20s" Dec 05 10:37:39.459094 master-0 kubenswrapper[10444]: I1205 10:37:39.456094 10444 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456099 10444 flags.go:64] FLAG: --image-credential-provider-config="" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456104 10444 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456110 10444 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456115 10444 flags.go:64] FLAG: --image-service-endpoint="" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456120 10444 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456127 10444 flags.go:64] FLAG: --kube-api-burst="100" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456132 10444 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456137 10444 flags.go:64] FLAG: --kube-api-qps="50" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456142 10444 flags.go:64] FLAG: --kube-reserved="" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456147 10444 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456152 10444 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456157 10444 flags.go:64] FLAG: --kubelet-cgroups="" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456162 10444 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456167 10444 flags.go:64] FLAG: --lock-file="" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456172 10444 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456176 10444 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456181 10444 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456190 10444 flags.go:64] FLAG: --log-json-split-stream="false" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456195 10444 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456200 10444 flags.go:64] FLAG: --log-text-split-stream="false" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456205 10444 flags.go:64] FLAG: --logging-format="text" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456210 10444 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456215 10444 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456221 10444 flags.go:64] FLAG: --manifest-url="" Dec 05 10:37:39.459885 master-0 kubenswrapper[10444]: I1205 10:37:39.456227 10444 flags.go:64] FLAG: --manifest-url-header="" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456234 10444 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456239 10444 flags.go:64] FLAG: --max-open-files="1000000" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456246 10444 flags.go:64] FLAG: --max-pods="110" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456255 10444 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456260 10444 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456265 10444 flags.go:64] FLAG: --memory-manager-policy="None" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456270 10444 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456275 10444 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456279 10444 flags.go:64] FLAG: --node-ip="192.168.32.10" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456285 10444 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456296 10444 flags.go:64] FLAG: --node-status-max-images="50" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456301 10444 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456306 10444 flags.go:64] FLAG: --oom-score-adj="-999" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456311 10444 flags.go:64] FLAG: --pod-cidr="" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456317 10444 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a70b2a95140d1e90978f36cc9889013ae34bd232662c5424002274385669ed9" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456326 10444 flags.go:64] FLAG: --pod-manifest-path="" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456330 10444 flags.go:64] FLAG: --pod-max-pids="-1" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456335 10444 flags.go:64] FLAG: --pods-per-core="0" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456340 10444 flags.go:64] FLAG: --port="10250" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456345 10444 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456350 10444 flags.go:64] FLAG: --provider-id="" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456355 10444 flags.go:64] FLAG: --qos-reserved="" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456359 10444 flags.go:64] FLAG: --read-only-port="10255" Dec 05 10:37:39.460676 master-0 kubenswrapper[10444]: I1205 10:37:39.456364 10444 flags.go:64] FLAG: --register-node="true" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456370 10444 flags.go:64] FLAG: --register-schedulable="true" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456374 10444 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456383 10444 flags.go:64] FLAG: --registry-burst="10" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456388 10444 flags.go:64] FLAG: --registry-qps="5" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456394 10444 flags.go:64] FLAG: --reserved-cpus="" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456399 10444 flags.go:64] FLAG: --reserved-memory="" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456408 10444 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456414 10444 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456423 10444 flags.go:64] FLAG: --rotate-certificates="false" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456428 10444 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456449 10444 flags.go:64] FLAG: --runonce="false" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456456 10444 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456464 10444 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456470 10444 flags.go:64] FLAG: --seccomp-default="false" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456475 10444 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456480 10444 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456485 10444 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456490 10444 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456495 10444 flags.go:64] FLAG: --storage-driver-password="root" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456500 10444 flags.go:64] FLAG: --storage-driver-secure="false" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456506 10444 flags.go:64] FLAG: --storage-driver-table="stats" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456511 10444 flags.go:64] FLAG: --storage-driver-user="root" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456516 10444 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456521 10444 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 05 10:37:39.461623 master-0 kubenswrapper[10444]: I1205 10:37:39.456527 10444 flags.go:64] FLAG: --system-cgroups="" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456532 10444 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456540 10444 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456545 10444 flags.go:64] FLAG: --tls-cert-file="" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456550 10444 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456556 10444 flags.go:64] FLAG: --tls-min-version="" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456561 10444 flags.go:64] FLAG: --tls-private-key-file="" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456566 10444 flags.go:64] FLAG: --topology-manager-policy="none" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456572 10444 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456577 10444 flags.go:64] FLAG: --topology-manager-scope="container" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456582 10444 flags.go:64] FLAG: --v="2" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456589 10444 flags.go:64] FLAG: --version="false" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456596 10444 flags.go:64] FLAG: --vmodule="" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456603 10444 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: I1205 10:37:39.456609 10444 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: W1205 10:37:39.456719 10444 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: W1205 10:37:39.456728 10444 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: W1205 10:37:39.456733 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: W1205 10:37:39.456738 10444 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: W1205 10:37:39.456743 10444 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: W1205 10:37:39.456751 10444 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: W1205 10:37:39.456757 10444 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: W1205 10:37:39.456762 10444 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:37:39.462860 master-0 kubenswrapper[10444]: W1205 10:37:39.456767 10444 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456772 10444 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456776 10444 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456781 10444 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456785 10444 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456789 10444 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456794 10444 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456799 10444 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456803 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456807 10444 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456811 10444 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456816 10444 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456824 10444 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456829 10444 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456833 10444 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456837 10444 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456843 10444 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456847 10444 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456852 10444 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456856 10444 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:37:39.463705 master-0 kubenswrapper[10444]: W1205 10:37:39.456860 10444 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456865 10444 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456869 10444 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456874 10444 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456878 10444 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456882 10444 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456887 10444 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456891 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456895 10444 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456903 10444 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456908 10444 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456912 10444 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456917 10444 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456921 10444 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456926 10444 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456930 10444 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456934 10444 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456938 10444 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456943 10444 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456947 10444 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:37:39.464275 master-0 kubenswrapper[10444]: W1205 10:37:39.456952 10444 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.456956 10444 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.456960 10444 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.456964 10444 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.456972 10444 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.456976 10444 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.456980 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.456985 10444 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.456990 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.456995 10444 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.457002 10444 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.457007 10444 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.457013 10444 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.457019 10444 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.457024 10444 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.457029 10444 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.457035 10444 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.457039 10444 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.457044 10444 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:37:39.464846 master-0 kubenswrapper[10444]: W1205 10:37:39.457049 10444 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: W1205 10:37:39.457054 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: W1205 10:37:39.457061 10444 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: W1205 10:37:39.457068 10444 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: W1205 10:37:39.457074 10444 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: I1205 10:37:39.457081 10444 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false StreamingCollectionEncodingToJSON:false StreamingCollectionEncodingToProtobuf:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: I1205 10:37:39.464739 10444 server.go:491] "Kubelet version" kubeletVersion="v1.31.13" Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: I1205 10:37:39.464773 10444 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: W1205 10:37:39.464855 10444 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: W1205 10:37:39.464863 10444 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: W1205 10:37:39.464871 10444 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: W1205 10:37:39.464877 10444 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: W1205 10:37:39.464884 10444 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: W1205 10:37:39.464891 10444 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:37:39.465700 master-0 kubenswrapper[10444]: W1205 10:37:39.464897 10444 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464902 10444 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464906 10444 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464911 10444 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464916 10444 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464920 10444 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464925 10444 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464930 10444 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464934 10444 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464940 10444 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464945 10444 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464949 10444 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464954 10444 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464958 10444 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464963 10444 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464967 10444 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464972 10444 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464977 10444 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464983 10444 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464990 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:37:39.466311 master-0 kubenswrapper[10444]: W1205 10:37:39.464997 10444 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465003 10444 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465008 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465013 10444 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465018 10444 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465024 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465029 10444 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465033 10444 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465038 10444 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465043 10444 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465047 10444 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465052 10444 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465056 10444 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465082 10444 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465091 10444 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465096 10444 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465102 10444 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465107 10444 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465113 10444 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465119 10444 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:37:39.467111 master-0 kubenswrapper[10444]: W1205 10:37:39.465124 10444 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465129 10444 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465134 10444 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465139 10444 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465145 10444 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465150 10444 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465155 10444 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465161 10444 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465165 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465170 10444 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465175 10444 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465180 10444 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465185 10444 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465190 10444 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465195 10444 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465200 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465204 10444 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465209 10444 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465213 10444 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465217 10444 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:37:39.469829 master-0 kubenswrapper[10444]: W1205 10:37:39.465225 10444 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465232 10444 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465237 10444 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465242 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465246 10444 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465253 10444 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: I1205 10:37:39.465261 10444 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false StreamingCollectionEncodingToJSON:false StreamingCollectionEncodingToProtobuf:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465401 10444 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465410 10444 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465415 10444 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465424 10444 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465430 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465452 10444 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465458 10444 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:37:39.470568 master-0 kubenswrapper[10444]: W1205 10:37:39.465463 10444 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465469 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465473 10444 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465479 10444 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465484 10444 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465489 10444 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465494 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465498 10444 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465502 10444 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465507 10444 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465513 10444 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465520 10444 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465526 10444 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465531 10444 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465537 10444 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465541 10444 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465546 10444 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465551 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465556 10444 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465561 10444 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:37:39.471874 master-0 kubenswrapper[10444]: W1205 10:37:39.465565 10444 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465569 10444 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465574 10444 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465579 10444 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465585 10444 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465591 10444 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465596 10444 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465601 10444 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465607 10444 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465612 10444 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465616 10444 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465622 10444 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465627 10444 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465632 10444 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465636 10444 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465642 10444 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465648 10444 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465655 10444 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465660 10444 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:37:39.473643 master-0 kubenswrapper[10444]: W1205 10:37:39.465665 10444 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465670 10444 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465675 10444 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465679 10444 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465684 10444 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465689 10444 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465694 10444 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465698 10444 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465703 10444 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465709 10444 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465714 10444 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465719 10444 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465725 10444 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465729 10444 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465734 10444 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465739 10444 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465743 10444 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465748 10444 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465752 10444 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465757 10444 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:37:39.474579 master-0 kubenswrapper[10444]: W1205 10:37:39.465761 10444 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: W1205 10:37:39.465766 10444 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: W1205 10:37:39.465770 10444 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: W1205 10:37:39.465775 10444 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: W1205 10:37:39.465780 10444 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: W1205 10:37:39.465784 10444 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: I1205 10:37:39.465793 10444 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false StreamingCollectionEncodingToJSON:false StreamingCollectionEncodingToProtobuf:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: I1205 10:37:39.465978 10444 server.go:940] "Client rotation is on, will bootstrap in background" Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: I1205 10:37:39.468073 10444 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: I1205 10:37:39.468166 10444 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: I1205 10:37:39.468472 10444 server.go:997] "Starting client certificate rotation" Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: I1205 10:37:39.468484 10444 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 05 10:37:39.475766 master-0 kubenswrapper[10444]: I1205 10:37:39.468679 10444 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2025-12-06 10:27:25 +0000 UTC, rotation deadline is 2025-12-06 07:16:45.621825819 +0000 UTC Dec 05 10:37:39.476341 master-0 kubenswrapper[10444]: I1205 10:37:39.468750 10444 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 20h39m6.15307926s for next certificate rotation Dec 05 10:37:39.476341 master-0 kubenswrapper[10444]: I1205 10:37:39.469167 10444 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 10:37:39.476341 master-0 kubenswrapper[10444]: I1205 10:37:39.470598 10444 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 10:37:39.476341 master-0 kubenswrapper[10444]: I1205 10:37:39.473548 10444 log.go:25] "Validated CRI v1 runtime API" Dec 05 10:37:39.476341 master-0 kubenswrapper[10444]: I1205 10:37:39.476100 10444 log.go:25] "Validated CRI v1 image API" Dec 05 10:37:39.477315 master-0 kubenswrapper[10444]: I1205 10:37:39.477271 10444 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 05 10:37:39.486627 master-0 kubenswrapper[10444]: I1205 10:37:39.486561 10444 fs.go:135] Filesystem UUIDs: map[354073b4-663b-4c82-b811-53676c745317:/dev/vda3 7B77-95E7:/dev/vda2 910678ff-f77e-4a7d-8d53-86f2ac47a823:/dev/vda4] Dec 05 10:37:39.487377 master-0 kubenswrapper[10444]: I1205 10:37:39.486611 10444 fs.go:136] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/09d6c752fa5b21d6de28695ca1e30252206036602e7b7cad7c82ecf0c820105d/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/09d6c752fa5b21d6de28695ca1e30252206036602e7b7cad7c82ecf0c820105d/userdata/shm major:0 minor:46 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/0d2211dcab91af9edc388c398c5fc7a90c2e8c65270b4ebc9dead574b96eaefe/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/0d2211dcab91af9edc388c398c5fc7a90c2e8c65270b4ebc9dead574b96eaefe/userdata/shm major:0 minor:332 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/12d79e8d025da0e5c89728d8b35230c8c9764259d39417d2ff38f00dc9c263f6/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/12d79e8d025da0e5c89728d8b35230c8c9764259d39417d2ff38f00dc9c263f6/userdata/shm major:0 minor:349 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/4833f667ac4451a150ed477789be72c08b65162d8c7bbf29e44cd656299dbd63/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/4833f667ac4451a150ed477789be72c08b65162d8c7bbf29e44cd656299dbd63/userdata/shm major:0 minor:375 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/5ab705ed0c0dfea5ee8845410f03f5084071123b3566b80a7ef7e6fcce0fa974/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/5ab705ed0c0dfea5ee8845410f03f5084071123b3566b80a7ef7e6fcce0fa974/userdata/shm major:0 minor:383 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/5b905a3a36c591b5f3704d14e85aafe47726c309b50985c9308b5ed0d4a90267/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/5b905a3a36c591b5f3704d14e85aafe47726c309b50985c9308b5ed0d4a90267/userdata/shm major:0 minor:346 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/611473a6bb698264f5594083fb6219f23f5848bd85b0fdd8dd5932b32fae3c36/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/611473a6bb698264f5594083fb6219f23f5848bd85b0fdd8dd5932b32fae3c36/userdata/shm major:0 minor:50 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/6607da6ec4214a3fe2d315e70d9756b15e987231bb8310517d849da88c8e7b15/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/6607da6ec4214a3fe2d315e70d9756b15e987231bb8310517d849da88c8e7b15/userdata/shm major:0 minor:165 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/6ead10efe7a643d9b0dd883cc19f8ef852a5658bc79c235b7c7c6b5de2e97811/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/6ead10efe7a643d9b0dd883cc19f8ef852a5658bc79c235b7c7c6b5de2e97811/userdata/shm major:0 minor:189 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/70e7f53d39e81bfad8819ec3c2244577ab77bdd0b9a5512ba3a7cb61ae8bafb4/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/70e7f53d39e81bfad8819ec3c2244577ab77bdd0b9a5512ba3a7cb61ae8bafb4/userdata/shm major:0 minor:473 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/7672f24a6f45e585db5dc690d30bd20c69d755aa3cc9d4346ea6ca7923305116/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/7672f24a6f45e585db5dc690d30bd20c69d755aa3cc9d4346ea6ca7923305116/userdata/shm major:0 minor:54 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/7e22b55ec9c49898cca9f69ea15552e8a230cbb6781fbf80f0e4919b22b7b7d8/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/7e22b55ec9c49898cca9f69ea15552e8a230cbb6781fbf80f0e4919b22b7b7d8/userdata/shm major:0 minor:338 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/8114256db0301425275a52e5c6f4316fb7a7d5eed5146413b166f9ab3918f143/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/8114256db0301425275a52e5c6f4316fb7a7d5eed5146413b166f9ab3918f143/userdata/shm major:0 minor:341 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/87b1749af9a101502a0aa9cd2d75f0e1562789f11a001117a5d269d9ed620723/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/87b1749af9a101502a0aa9cd2d75f0e1562789f11a001117a5d269d9ed620723/userdata/shm major:0 minor:58 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/89c8e517113e00942d58576e87acb9fd6a3e8f726e30d612913a3d827c50d10f/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/89c8e517113e00942d58576e87acb9fd6a3e8f726e30d612913a3d827c50d10f/userdata/shm major:0 minor:485 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/8d62f303e35a44bd684cd9b7f51835e4e80959496c433fcc66b39cdd84c84cc8/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/8d62f303e35a44bd684cd9b7f51835e4e80959496c433fcc66b39cdd84c84cc8/userdata/shm major:0 minor:149 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/ad7983dc1d046a27e0480393c3ce354f96d0968b079cf24fa1ebf196a469d66d/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/ad7983dc1d046a27e0480393c3ce354f96d0968b079cf24fa1ebf196a469d66d/userdata/shm major:0 minor:127 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/b44ac27d1b7f531e11c3f4d352d89c4a7679a795ec64417e43c87c5d1e725e90/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/b44ac27d1b7f531e11c3f4d352d89c4a7679a795ec64417e43c87c5d1e725e90/userdata/shm major:0 minor:314 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/b7452830de99c0487a8d247b23da92391c3a1113eb379e2b40ac1673b8cb1fe1/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/b7452830de99c0487a8d247b23da92391c3a1113eb379e2b40ac1673b8cb1fe1/userdata/shm major:0 minor:336 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/bdfbd9daf37e036f9b2d20314905a9aa48f68e21954c4e133ee980fe10b0d8d2/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/bdfbd9daf37e036f9b2d20314905a9aa48f68e21954c4e133ee980fe10b0d8d2/userdata/shm major:0 minor:343 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/ccbec542fddb9bd0831349589c4ac65650d79b1cf75a0af75cf8ca2d1a0d89f6/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/ccbec542fddb9bd0831349589c4ac65650d79b1cf75a0af75cf8ca2d1a0d89f6/userdata/shm major:0 minor:337 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7/userdata/shm major:0 minor:41 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/e2506cc6607d77e49135fcb375a37bafe2d511f6ed75718677873e40b092ab87/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/e2506cc6607d77e49135fcb375a37bafe2d511f6ed75718677873e40b092ab87/userdata/shm major:0 minor:333 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/eecf6d3a967ce077414ad07bec5a21f5acf63791b8f14de4d8d34c6cfe91444a/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/eecf6d3a967ce077414ad07bec5a21f5acf63791b8f14de4d8d34c6cfe91444a/userdata/shm major:0 minor:468 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/efc2234f689663f4f73596bfe5f7a66235fb095a2337fa0c8bc412e1e08433fb/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/efc2234f689663f4f73596bfe5f7a66235fb095a2337fa0c8bc412e1e08433fb/userdata/shm major:0 minor:139 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/fc75409854b1370368a39cd91bcd5c4ac2aa319c2a8b9aaad1e7abab3fd56a49/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/fc75409854b1370368a39cd91bcd5c4ac2aa319c2a8b9aaad1e7abab3fd56a49/userdata/shm major:0 minor:179 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/fedfe3c8bf160c89d4d691575ab31b377473c05a2dd753786d2ef1a80ea5d562/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/fedfe3c8bf160c89d4d691575ab31b377473c05a2dd753786d2ef1a80ea5d562/userdata/shm major:0 minor:347 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~projected/kube-api-access:{mountpoint:/var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~projected/kube-api-access major:0 minor:281 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~secret/serving-cert major:0 minor:256 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~projected/kube-api-access-cwqkb:{mountpoint:/var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~projected/kube-api-access-cwqkb major:0 minor:122 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~secret/metrics-tls:{mountpoint:/var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~secret/metrics-tls major:0 minor:43 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~projected/kube-api-access-7wfsv:{mountpoint:/var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~projected/kube-api-access-7wfsv major:0 minor:282 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~secret/serving-cert major:0 minor:257 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/bound-sa-token:{mountpoint:/var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/bound-sa-token major:0 minor:308 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/kube-api-access-vkqz7:{mountpoint:/var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/kube-api-access-vkqz7 major:0 minor:287 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/283122ba-be1c-4516-bd0f-df41c13c098b/volumes/kubernetes.io~projected/kube-api-access-vhjhk:{mountpoint:/var/lib/kubelet/pods/283122ba-be1c-4516-bd0f-df41c13c098b/volumes/kubernetes.io~projected/kube-api-access-vhjhk major:0 minor:275 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8/volumes/kubernetes.io~projected/kube-api-access-jrdsv:{mountpoint:/var/lib/kubelet/pods/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8/volumes/kubernetes.io~projected/kube-api-access-jrdsv major:0 minor:138 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c/volumes/kubernetes.io~projected/kube-api-access-ssssf:{mountpoint:/var/lib/kubelet/pods/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c/volumes/kubernetes.io~projected/kube-api-access-ssssf major:0 minor:153 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a/volumes/kubernetes.io~projected/kube-api-access-pdsh9:{mountpoint:/var/lib/kubelet/pods/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a/volumes/kubernetes.io~projected/kube-api-access-pdsh9 major:0 minor:472 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~projected/kube-api-access:{mountpoint:/var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~projected/kube-api-access major:0 minor:288 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~secret/serving-cert major:0 minor:260 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~projected/kube-api-access-jv4zs:{mountpoint:/var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~projected/kube-api-access-jv4zs major:0 minor:286 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~secret/profile-collector-cert:{mountpoint:/var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~secret/profile-collector-cert major:0 minor:254 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~projected/kube-api-access:{mountpoint:/var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~projected/kube-api-access major:0 minor:290 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~secret/serving-cert major:0 minor:258 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~projected/kube-api-access-6vp9b:{mountpoint:/var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~projected/kube-api-access-6vp9b major:0 minor:289 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~secret/cluster-olm-operator-serving-cert:{mountpoint:/var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~secret/cluster-olm-operator-serving-cert major:0 minor:263 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volume-subpaths/run-systemd/ovnkube-controller/6:{mountpoint:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volume-subpaths/run-systemd/ovnkube-controller/6 major:0 minor:24 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~projected/kube-api-access-tjgc4:{mountpoint:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~projected/kube-api-access-tjgc4 major:0 minor:172 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~secret/ovn-node-metrics-cert:{mountpoint:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~secret/ovn-node-metrics-cert major:0 minor:171 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/5a1bdc70-6412-47e0-8330-04d796cc8d55/volumes/kubernetes.io~projected/kube-api-access-5dtfn:{mountpoint:/var/lib/kubelet/pods/5a1bdc70-6412-47e0-8330-04d796cc8d55/volumes/kubernetes.io~projected/kube-api-access-5dtfn major:0 minor:292 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/6a9f011f-36f1-4308-a365-69425c186c7f/volumes/kubernetes.io~projected/kube-api-access-mvnxf:{mountpoint:/var/lib/kubelet/pods/6a9f011f-36f1-4308-a365-69425c186c7f/volumes/kubernetes.io~projected/kube-api-access-mvnxf major:0 minor:137 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes.io~projected/kube-api-access-fzwfq:{mountpoint:/var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes.io~projected/kube-api-access-fzwfq major:0 minor:266 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes.io~secret/serving-cert major:0 minor:264 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/720a7467-ce93-4d48-82ec-9ad0922d99c2/volumes/kubernetes.io~projected/kube-api-access-5n7tf:{mountpoint:/var/lib/kubelet/pods/720a7467-ce93-4d48-82ec-9ad0922d99c2/volumes/kubernetes.io~projected/kube-api-access-5n7tf major:0 minor:482 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/82ef99d4-41b7-4859-a5a1-4e31ce614a2a/volumes/kubernetes.io~projected/kube-api-access:{mountpoint:/var/lib/kubelet/pods/82ef99d4-41b7-4859-a5a1-4e31ce614a2a/volumes/kubernetes.io~projected/kube-api-access major:0 minor:77 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/87909f47-f2d7-46f8-a1c8-27336cdcce5d/volumes/kubernetes.io~projected/kube-api-access-rjrgm:{mountpoint:/var/lib/kubelet/pods/87909f47-f2d7-46f8-a1c8-27336cdcce5d/volumes/kubernetes.io~projected/kube-api-access-rjrgm major:0 minor:323 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/8c649a16-c187-412e-b5da-62a00bee38ab/volumes/kubernetes.io~projected/kube-api-access-d8pv2:{mountpoint:/var/lib/kubelet/pods/8c649a16-c187-412e-b5da-62a00bee38ab/volumes/kubernetes.io~projected/kube-api-access-d8pv2 major:0 minor:309 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/8d76404b-6d62-4a61-b6f6-0c8073eba198/volumes/kubernetes.io~projected/kube-api-access-7kb5c:{mountpoint:/var/lib/kubelet/pods/8d76404b-6d62-4a61-b6f6-0c8073eba198/volumes/kubernetes.io~projected/kube-api-access-7kb5c major:0 minor:293 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/926263c4-ec5b-41cb-9c30-0c88f636035f/volumes/kubernetes.io~projected/kube-api-access-fncz7:{mountpoint:/var/lib/kubelet/pods/926263c4-ec5b-41cb-9c30-0c88f636035f/volumes/kubernetes.io~projected/kube-api-access-fncz7 major:0 minor:318 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/a722cda9-29a0-4b7f-8e1d-9a8950ed765a/volumes/kubernetes.io~projected/kube-api-access-2lwgq:{mountpoint:/var/lib/kubelet/pods/a722cda9-29a0-4b7f-8e1d-9a8950ed765a/volumes/kubernetes.io~projected/kube-api-access-2lwgq major:0 minor:276 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~projected/kube-api-access-jfz87:{mountpoint:/var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~projected/kube-api-access-jfz87 major:0 minor:164 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~secret/ovn-control-plane-metrics-cert:{mountpoint:/var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~secret/ovn-control-plane-metrics-cert major:0 minor:163 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~projected/kube-api-access-98qt8:{mountpoint:/var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~projected/kube-api-access-98qt8 major:0 minor:273 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~secret/serving-cert major:0 minor:259 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~projected/kube-api-access-fkmvj:{mountpoint:/var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~projected/kube-api-access-fkmvj major:0 minor:274 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~secret/serving-cert major:0 minor:265 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~projected/kube-api-access-wg9zq:{mountpoint:/var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~projected/kube-api-access-wg9zq major:0 minor:188 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~secret/webhook-cert:{mountpoint:/var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~secret/webhook-cert major:0 minor:187 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~projected/kube-api-access-2fbss:{mountpoint:/var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~projected/kube-api-access-2fbss major:0 minor:313 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~secret/serving-cert major:0 minor:262 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/volumes/kubernetes.io~projected/kube-api-access-ml74w:{mountpoint:/var/lib/kubelet/pods/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/volumes/kubernetes.io~projected/kube-api-access-ml74w major:0 minor:465 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~projected/kube-api-access-qn7rj:{mountpoint:/var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~projected/kube-api-access-qn7rj major:0 minor:267 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~projected/kube-api-access-twh85:{mountpoint:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~projected/kube-api-access-twh85 major:0 minor:299 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/etcd-client:{mountpoint:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/etcd-client major:0 minor:261 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/serving-cert major:0 minor:255 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/bound-sa-token:{mountpoint:/var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/bound-sa-token major:0 minor:312 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/kube-api-access-khfxz:{mountpoint:/var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/kube-api-access-khfxz major:0 minor:296 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~projected/kube-api-access-h2vd4:{mountpoint:/var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~projected/kube-api-access-h2vd4 major:0 minor:325 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~secret/serving-cert major:0 minor:277 fsType:tmpfs blockSize:0} overlay_0-104:{mountpoint:/var/lib/containers/storage/overlay/a20d953fc8dd242399207d713ec1054f28c895f42ea580b06460d05cffa3251a/merged major:0 minor:104 fsType:overlay blockSize:0} overlay_0-112:{mountpoint:/var/lib/containers/storage/overlay/a1a1563194f27a58be2083d6f9552ed4fc07b961f872c80dc02a577439af6444/merged major:0 minor:112 fsType:overlay blockSize:0} overlay_0-120:{mountpoint:/var/lib/containers/storage/overlay/d1e62255e12c370b98430fede60367d62bee912ad46005c4d932d9c04e44d755/merged major:0 minor:120 fsType:overlay blockSize:0} overlay_0-129:{mountpoint:/var/lib/containers/storage/overlay/69d3b41aab4f992addc9aa3d6f15431647883b3c252fe16c21c0b872baf1204f/merged major:0 minor:129 fsType:overlay blockSize:0} overlay_0-131:{mountpoint:/var/lib/containers/storage/overlay/71e74b988ee1323c79a5fc15ad72123627de924888792d10c63849dd96a813c1/merged major:0 minor:131 fsType:overlay blockSize:0} overlay_0-133:{mountpoint:/var/lib/containers/storage/overlay/daacf88f373af098aafa20a9a3771ec853b70408688f81c00b21bb58d99a4f73/merged major:0 minor:133 fsType:overlay blockSize:0} overlay_0-135:{mountpoint:/var/lib/containers/storage/overlay/5726ddc4394935b309cbcd1f78eff930a45de2988789cba3b8ad5ad40a5b2d87/merged major:0 minor:135 fsType:overlay blockSize:0} overlay_0-147:{mountpoint:/var/lib/containers/storage/overlay/b0b75e0c84c872067c1cc7c0120ef370dfc2f8cb5b8b502f3235970ed83eb3c3/merged major:0 minor:147 fsType:overlay blockSize:0} overlay_0-151:{mountpoint:/var/lib/containers/storage/overlay/ad73c9feb4b6df5992086ded295e9447fe56551a6bc2abfe15c347ec252b2df2/merged major:0 minor:151 fsType:overlay blockSize:0} overlay_0-154:{mountpoint:/var/lib/containers/storage/overlay/ecb987277e86a099a527f281a809eca6a1e61d1361d335422ef901559beeac23/merged major:0 minor:154 fsType:overlay blockSize:0} overlay_0-156:{mountpoint:/var/lib/containers/storage/overlay/2adaf01a0b08d13626a4da6df7e18b5e4907881442b28b7d90d8833f5c3ac5e8/merged major:0 minor:156 fsType:overlay blockSize:0} overlay_0-158:{mountpoint:/var/lib/containers/storage/overlay/8042c20fe3cb599e0d51ab721f80c316642c0ca72e8c9a9fc7cec85eeec21221/merged major:0 minor:158 fsType:overlay blockSize:0} overlay_0-167:{mountpoint:/var/lib/containers/storage/overlay/221c2f7e66b93d86e9da7232d411202d0694cf12b29cb712c4960a5e5e5c5024/merged major:0 minor:167 fsType:overlay blockSize:0} overlay_0-169:{mountpoint:/var/lib/containers/storage/overlay/c1f05399228e24704e4f3567767668c5f8597f886914f28b2747fe881ccf7121/merged major:0 minor:169 fsType:overlay blockSize:0} overlay_0-181:{mountpoint:/var/lib/containers/storage/overlay/c9fc3ab72976c7274a75cabc89076c5a3c76f05b2459d997d644b3fb0157cc5d/merged major:0 minor:181 fsType:overlay blockSize:0} overlay_0-183:{mountpoint:/var/lib/containers/storage/overlay/c8112fdb855c50fd639b9394dc2647a4c50f459f0efe7945c1ad234068025570/merged major:0 minor:183 fsType:overlay blockSize:0} overlay_0-185:{mountpoint:/var/lib/containers/storage/overlay/cd4f4481b0db7e5c34b29bfc60db4cceaf613a8470ee3c3c0eb9b6f37f70f133/merged major:0 minor:185 fsType:overlay blockSize:0} overlay_0-191:{mountpoint:/var/lib/containers/storage/overlay/672ba5764d28fa61ecdf4fa2dedbb811dd695a512500f55ce22fb456e4aaea22/merged major:0 minor:191 fsType:overlay blockSize:0} overlay_0-193:{mountpoint:/var/lib/containers/storage/overlay/44634517510c9242ca7a2ef11d48400d2344fe1b871f8b485681cfc8c11a0f75/merged major:0 minor:193 fsType:overlay blockSize:0} overlay_0-195:{mountpoint:/var/lib/containers/storage/overlay/d98934f7894d10a0de0cf8eaa27f3e1cf20717558b2da855607a3157b9a43913/merged major:0 minor:195 fsType:overlay blockSize:0} overlay_0-197:{mountpoint:/var/lib/containers/storage/overlay/0b6fc7c9d4ce32891e9527036aec70da87ba1aab65c0ee4ccd9d0a818a4fedb1/merged major:0 minor:197 fsType:overlay blockSize:0} overlay_0-199:{mountpoint:/var/lib/containers/storage/overlay/c700fbca17bf0aa18390fa0e7a3e7e50c95407e97b16040c11ff75c145e3024b/merged major:0 minor:199 fsType:overlay blockSize:0} overlay_0-205:{mountpoint:/var/lib/containers/storage/overlay/ee131bae2ad4ad92689b394d63362df84d3c96ef9aa039e40b427edb62c1d215/merged major:0 minor:205 fsType:overlay blockSize:0} overlay_0-209:{mountpoint:/var/lib/containers/storage/overlay/739cb662179f4420ea4e414cb5ccb3a85f1ebfdd8b84bd61ef7f9ddc35dc8308/merged major:0 minor:209 fsType:overlay blockSize:0} overlay_0-213:{mountpoint:/var/lib/containers/storage/overlay/746dae4fca92f1f46856320f3142b191f82009bfc78da76cf9934fe6c479eb21/merged major:0 minor:213 fsType:overlay blockSize:0} overlay_0-228:{mountpoint:/var/lib/containers/storage/overlay/a0176a28a70d267f880769f37ef654ae085627da7e2a26813af8a468a87ad60f/merged major:0 minor:228 fsType:overlay blockSize:0} overlay_0-236:{mountpoint:/var/lib/containers/storage/overlay/f77b16658281bdf8a111ea497129031d7877cb11785b94d3092cb391f2202ca5/merged major:0 minor:236 fsType:overlay blockSize:0} overlay_0-244:{mountpoint:/var/lib/containers/storage/overlay/43ebf08663fa907da55be47d20f8b55b612ac6419332c76f1b701e88c5a677ae/merged major:0 minor:244 fsType:overlay blockSize:0} overlay_0-252:{mountpoint:/var/lib/containers/storage/overlay/99102a8dcf874c4ba63e522e1b531e1c54ff115e500f3b8f04ae20ce91113752/merged major:0 minor:252 fsType:overlay blockSize:0} overlay_0-268:{mountpoint:/var/lib/containers/storage/overlay/f11656f01663f1840822c3d5e2d7697ce24b7c255ab7bb1c54a68c7e6d64874b/merged major:0 minor:268 fsType:overlay blockSize:0} overlay_0-291:{mountpoint:/var/lib/containers/storage/overlay/c66d86781bfb63b46591c93693ca7996ba6b846502222a9ccb6778144909ebce/merged major:0 minor:291 fsType:overlay blockSize:0} overlay_0-294:{mountpoint:/var/lib/containers/storage/overlay/e8443993f3b7416ea6178a6b91bc11a40a839a0b6dd6475068141f140701c195/merged major:0 minor:294 fsType:overlay blockSize:0} overlay_0-316:{mountpoint:/var/lib/containers/storage/overlay/9e2277d70b2152dc3fcea05e7c7c9a5e5c37d900f125f6c37e400df22fa7a95b/merged major:0 minor:316 fsType:overlay blockSize:0} overlay_0-320:{mountpoint:/var/lib/containers/storage/overlay/6082bb1f87abff06cf3c105bba94e1ee8b30bed4bc88f2a11299cd54f691aa94/merged major:0 minor:320 fsType:overlay blockSize:0} overlay_0-322:{mountpoint:/var/lib/containers/storage/overlay/89b22bc61d954b1a593c909cbfe0436e763a79a3c3a22177c93dea7245201f66/merged major:0 minor:322 fsType:overlay blockSize:0} overlay_0-352:{mountpoint:/var/lib/containers/storage/overlay/07140a77fa648deb8b4d88c22e68fa10f6c7bac30b0c7786d3185b8459aeb839/merged major:0 minor:352 fsType:overlay blockSize:0} overlay_0-354:{mountpoint:/var/lib/containers/storage/overlay/49d8184721481ae0b07ee14398595a506ba2aae20f878baef545d290170ae9f1/merged major:0 minor:354 fsType:overlay blockSize:0} overlay_0-356:{mountpoint:/var/lib/containers/storage/overlay/0ccd97d1f30e293c96fca5436504b3210f4918b32e09318657425e960f17bc06/merged major:0 minor:356 fsType:overlay blockSize:0} overlay_0-358:{mountpoint:/var/lib/containers/storage/overlay/58b301a47cec9284ac1182393bdb11ccde7126768e1617ddb5f47b35e855e4dc/merged major:0 minor:358 fsType:overlay blockSize:0} overlay_0-360:{mountpoint:/var/lib/containers/storage/overlay/8c068bfa8aefed5fa7a6eaf9f0e7935b02af9100c42db54f0a8366f55dd75917/merged major:0 minor:360 fsType:overlay blockSize:0} overlay_0-362:{mountpoint:/var/lib/containers/storage/overlay/baeecbd481f89e1f70bbbe28bba0e050b3fe21bdfed072b3fc7cabdae408ba54/merged major:0 minor:362 fsType:overlay blockSize:0} overlay_0-364:{mountpoint:/var/lib/containers/storage/overlay/77eafadc8613b6f7fa1dd04e077d6685de20f2833bd5d2ecd5481f16ed6776cb/merged major:0 minor:364 fsType:overlay blockSize:0} overlay_0-366:{mountpoint:/var/lib/containers/storage/overlay/0b5b6e1e1402a79ebc6057fff4d379db33ef577dd917b6064f865035fa92d27e/merged major:0 minor:366 fsType:overlay blockSize:0} overlay_0-377:{mountpoint:/var/lib/containers/storage/overlay/15945a606966e8a5108e79db1845d2e9f5c51d24e304a2b03a319a76ced49309/merged major:0 minor:377 fsType:overlay blockSize:0} overlay_0-379:{mountpoint:/var/lib/containers/storage/overlay/d75c2056c63f1f756f7f527f91b3e94a861db48794706dd5523843f602ef7564/merged major:0 minor:379 fsType:overlay blockSize:0} overlay_0-381:{mountpoint:/var/lib/containers/storage/overlay/7246059058bd272c6b057dd7db6a7932b3512a067942e59be4995fcad774d800/merged major:0 minor:381 fsType:overlay blockSize:0} overlay_0-385:{mountpoint:/var/lib/containers/storage/overlay/d11123e76f1759aec07fd01ca6d1cd7946e7ce3359a0b20ed6518004832ddf75/merged major:0 minor:385 fsType:overlay blockSize:0} overlay_0-387:{mountpoint:/var/lib/containers/storage/overlay/855cc47dac6e9aeff40befed964a58516a945a51ec4b822c26f9f7e9ade252e5/merged major:0 minor:387 fsType:overlay blockSize:0} overlay_0-390:{mountpoint:/var/lib/containers/storage/overlay/918e0f320e7363ef658bcec35c9f6378d9f7164e94afa7773a4528ab3dac35e9/merged major:0 minor:390 fsType:overlay blockSize:0} overlay_0-392:{mountpoint:/var/lib/containers/storage/overlay/742c9fd3579942248a2b4f159fbe5a99b1e051eb9ff0418d64ab36c59b6a77b3/merged major:0 minor:392 fsType:overlay blockSize:0} overlay_0-394:{mountpoint:/var/lib/containers/storage/overlay/bb5f8d74311e59f2fad895aed10893a4e941090f783a1b361a7bea909dc56351/merged major:0 minor:394 fsType:overlay blockSize:0} overlay_0-396:{mountpoint:/var/lib/containers/storage/overlay/e8bcc5489be28aaa95c515c855d8d50a44833d2ba1061d4caef85995e2c8fa22/merged major:0 minor:396 fsType:overlay blockSize:0} overlay_0-398:{mountpoint:/var/lib/containers/storage/overlay/9a30646052631ad7471a38651b056f6e4bfd683684942d114a770b463630e758/merged major:0 minor:398 fsType:overlay blockSize:0} overlay_0-400:{mountpoint:/var/lib/containers/storage/overlay/dc01f2c13de248e69d5c1a228ac808833202ed8e4d3f6182e161bcb146fb7f9c/merged major:0 minor:400 fsType:overlay blockSize:0} overlay_0-402:{mountpoint:/var/lib/containers/storage/overlay/163f2360566f2d7fa1a96361b76bd2218f5bd18089cca6c097c5b6a9e5d3f981/merged major:0 minor:402 fsType:overlay blockSize:0} overlay_0-404:{mountpoint:/var/lib/containers/storage/overlay/d08dd47ff25650be848a4aa97e3cde30bcf985640a213d05a8039f1b4ca3be9f/merged major:0 minor:404 fsType:overlay blockSize:0} overlay_0-406:{mountpoint:/var/lib/containers/storage/overlay/78b1844a69b54e740544612081f919ebd2d077f51b219b4ec0638fae6ed26ffe/merged major:0 minor:406 fsType:overlay blockSize:0} overlay_0-408:{mountpoint:/var/lib/containers/storage/overlay/9ac49532fb6a348bdb2072a72871b39f8bd086cf65f09fdeea58728b5ae29776/merged major:0 minor:408 fsType:overlay blockSize:0} overlay_0-410:{mountpoint:/var/lib/containers/storage/overlay/5d437a71bdb0c756d10ba5b0455fff4ada03aaddb2a941ef4e974dd41bbf32b2/merged major:0 minor:410 fsType:overlay blockSize:0} overlay_0-44:{mountpoint:/var/lib/containers/storage/overlay/3c305c53cfd29952e6ca5d5fab57f55943a07ccb4b003a5beff65a97cd1655a4/merged major:0 minor:44 fsType:overlay blockSize:0} overlay_0-470:{mountpoint:/var/lib/containers/storage/overlay/39a618abcf5892e92a5c31e18efea6b29947cdf15eab61d5e1cc4c561da7187c/merged major:0 minor:470 fsType:overlay blockSize:0} overlay_0-48:{mountpoint:/var/lib/containers/storage/overlay/5134969f743a5c43e13d232355de02f664c9276e9c337c027efc5acad4394e09/merged major:0 minor:48 fsType:overlay blockSize:0} overlay_0-483:{mountpoint:/var/lib/containers/storage/overlay/9df50ee91ad13587e9da1d293b4056606b333ac5520288725c0864b2edb04538/merged major:0 minor:483 fsType:overlay blockSize:0} overlay_0-487:{mountpoint:/var/lib/containers/storage/overlay/da5c0507195eb3e459338c8f665b1b3f22beeab0a3a6187f6685895a0341a59c/merged major:0 minor:487 fsType:overlay blockSize:0} overlay_0-489:{mountpoint:/var/lib/containers/storage/overlay/0aa6590d30d44b4f2ac74813fd65513ddeef7027acf08afcd1172a45510f2d20/merged major:0 minor:489 fsType:overlay blockSize:0} overlay_0-52:{mountpoint:/var/lib/containers/storage/overlay/5a0332a00a0d1d2dabce939cc448ca4b410efd78a105f4d2943e3f6b73cf52d5/merged major:0 minor:52 fsType:overlay blockSize:0} overlay_0-56:{mountpoint:/var/lib/containers/storage/overlay/ea0fd6740244f2815be942e30415cd18b16fb7296be5ef4f321d9efeb4be2bd7/merged major:0 minor:56 fsType:overlay blockSize:0} overlay_0-60:{mountpoint:/var/lib/containers/storage/overlay/c9c40bece4e05d84e9b5cf4948ac7da1edb5fdea06c7e90158fe803b09aa7bee/merged major:0 minor:60 fsType:overlay blockSize:0} overlay_0-62:{mountpoint:/var/lib/containers/storage/overlay/e53497d94773e663d8dec87fb79f8ea0aa7dd58034c7327f49d504187e4bfacc/merged major:0 minor:62 fsType:overlay blockSize:0} overlay_0-64:{mountpoint:/var/lib/containers/storage/overlay/d2a7c15af1ff4298a7d24408e046de858446ae75f294a79121c199a388ae1922/merged major:0 minor:64 fsType:overlay blockSize:0} overlay_0-66:{mountpoint:/var/lib/containers/storage/overlay/5404c3d754818c42d503489d4cf4658afee5c26d898aac465526748f00c9879d/merged major:0 minor:66 fsType:overlay blockSize:0} overlay_0-74:{mountpoint:/var/lib/containers/storage/overlay/e3cfc7c2263909f5ccf44a734ed65e949425adebd5dc3a356d0cc54e77dccdbc/merged major:0 minor:74 fsType:overlay blockSize:0} overlay_0-76:{mountpoint:/var/lib/containers/storage/overlay/a3336bc13758cef0a6e91cbd627f2d88b4568114095e1579405b43a0487c668f/merged major:0 minor:76 fsType:overlay blockSize:0} overlay_0-78:{mountpoint:/var/lib/containers/storage/overlay/944162e22056c88fe51e6958f133e75eec6c33d01b43b3234e856fa6d99b0c74/merged major:0 minor:78 fsType:overlay blockSize:0} overlay_0-80:{mountpoint:/var/lib/containers/storage/overlay/ccdcec7c48fe79e9c8e76ea81516243a4f3eb9635142cd0ac87065086207b93c/merged major:0 minor:80 fsType:overlay blockSize:0} overlay_0-82:{mountpoint:/var/lib/containers/storage/overlay/286dfac26bc98a02688cf79358db73a0e7f0ab13a3315914649f939cc9e50b59/merged major:0 minor:82 fsType:overlay blockSize:0} overlay_0-90:{mountpoint:/var/lib/containers/storage/overlay/96bd0704ce5b27b4beb1ed44499a13f66e1aaae08ae1c65e19364b752727a87f/merged major:0 minor:90 fsType:overlay blockSize:0}] Dec 05 10:37:39.522338 master-0 kubenswrapper[10444]: I1205 10:37:39.521639 10444 manager.go:217] Machine: {Timestamp:2025-12-05 10:37:39.52039065 +0000 UTC m=+0.112303237 CPUVendorID:AuthenticAMD NumCores:16 NumPhysicalCores:1 NumSockets:16 CpuFrequency:2799998 MemoryCapacity:50514153472 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:9319397e1208473fb2e5a8b942c60cc5 SystemUUID:9319397e-1208-473f-b2e5-a8b942c60cc5 BootID:759865f8-012e-4147-b61d-8ab8adc2639f Filesystems:[{Device:/run/containers/storage/overlay-containers/8d62f303e35a44bd684cd9b7f51835e4e80959496c433fcc66b39cdd84c84cc8/userdata/shm DeviceMajor:0 DeviceMinor:149 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~secret/webhook-cert DeviceMajor:0 DeviceMinor:187 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:257 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:259 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/bound-sa-token DeviceMajor:0 DeviceMinor:308 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/8c649a16-c187-412e-b5da-62a00bee38ab/volumes/kubernetes.io~projected/kube-api-access-d8pv2 DeviceMajor:0 DeviceMinor:309 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/ad7983dc1d046a27e0480393c3ce354f96d0968b079cf24fa1ebf196a469d66d/userdata/shm DeviceMajor:0 DeviceMinor:127 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/8114256db0301425275a52e5c6f4316fb7a7d5eed5146413b166f9ab3918f143/userdata/shm DeviceMajor:0 DeviceMinor:341 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-410 DeviceMajor:0 DeviceMinor:410 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~projected/kube-api-access-98qt8 DeviceMajor:0 DeviceMinor:273 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-385 DeviceMajor:0 DeviceMinor:385 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7/userdata/shm DeviceMajor:0 DeviceMinor:41 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-158 DeviceMajor:0 DeviceMinor:158 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-213 DeviceMajor:0 DeviceMinor:213 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~projected/kube-api-access-tjgc4 DeviceMajor:0 DeviceMinor:172 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:256 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/a722cda9-29a0-4b7f-8e1d-9a8950ed765a/volumes/kubernetes.io~projected/kube-api-access-2lwgq DeviceMajor:0 DeviceMinor:276 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/87909f47-f2d7-46f8-a1c8-27336cdcce5d/volumes/kubernetes.io~projected/kube-api-access-rjrgm DeviceMajor:0 DeviceMinor:323 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/09d6c752fa5b21d6de28695ca1e30252206036602e7b7cad7c82ecf0c820105d/userdata/shm DeviceMajor:0 DeviceMinor:46 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/82ef99d4-41b7-4859-a5a1-4e31ce614a2a/volumes/kubernetes.io~projected/kube-api-access DeviceMajor:0 DeviceMinor:77 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-291 DeviceMajor:0 DeviceMinor:291 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/efc2234f689663f4f73596bfe5f7a66235fb095a2337fa0c8bc412e1e08433fb/userdata/shm DeviceMajor:0 DeviceMinor:139 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~secret/ovn-control-plane-metrics-cert DeviceMajor:0 DeviceMinor:163 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/kube-api-access-vkqz7 DeviceMajor:0 DeviceMinor:287 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/volumes/kubernetes.io~projected/kube-api-access-ml74w DeviceMajor:0 DeviceMinor:465 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-76 DeviceMajor:0 DeviceMinor:76 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c/volumes/kubernetes.io~projected/kube-api-access-ssssf DeviceMajor:0 DeviceMinor:153 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-379 DeviceMajor:0 DeviceMinor:379 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-487 DeviceMajor:0 DeviceMinor:487 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-52 DeviceMajor:0 DeviceMinor:52 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~projected/kube-api-access-6vp9b DeviceMajor:0 DeviceMinor:289 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~projected/kube-api-access-fkmvj DeviceMajor:0 DeviceMinor:274 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-316 DeviceMajor:0 DeviceMinor:316 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-366 DeviceMajor:0 DeviceMinor:366 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/720a7467-ce93-4d48-82ec-9ad0922d99c2/volumes/kubernetes.io~projected/kube-api-access-5n7tf DeviceMajor:0 DeviceMinor:482 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-66 DeviceMajor:0 DeviceMinor:66 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~secret/profile-collector-cert DeviceMajor:0 DeviceMinor:254 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:265 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/5b905a3a36c591b5f3704d14e85aafe47726c309b50985c9308b5ed0d4a90267/userdata/shm DeviceMajor:0 DeviceMinor:346 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-360 DeviceMajor:0 DeviceMinor:360 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-56 DeviceMajor:0 DeviceMinor:56 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-74 DeviceMajor:0 DeviceMinor:74 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-396 DeviceMajor:0 DeviceMinor:396 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/kube-api-access-khfxz DeviceMajor:0 DeviceMinor:296 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~projected/kube-api-access-h2vd4 DeviceMajor:0 DeviceMinor:325 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/b7452830de99c0487a8d247b23da92391c3a1113eb379e2b40ac1673b8cb1fe1/userdata/shm DeviceMajor:0 DeviceMinor:336 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-352 DeviceMajor:0 DeviceMinor:352 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:10102833152 Type:vfs Inodes:819200 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:25257078784 Type:vfs Inodes:1048576 HasInodes:true} {Device:overlay_0-205 DeviceMajor:0 DeviceMinor:205 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~projected/kube-api-access-7wfsv DeviceMajor:0 DeviceMinor:282 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volume-subpaths/run-systemd/ovnkube-controller/6 DeviceMajor:0 DeviceMinor:24 Capacity:10102833152 Type:vfs Inodes:819200 HasInodes:true} {Device:/run/containers/storage/overlay-containers/bdfbd9daf37e036f9b2d20314905a9aa48f68e21954c4e133ee980fe10b0d8d2/userdata/shm DeviceMajor:0 DeviceMinor:343 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/fedfe3c8bf160c89d4d691575ab31b377473c05a2dd753786d2ef1a80ea5d562/userdata/shm DeviceMajor:0 DeviceMinor:347 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/70e7f53d39e81bfad8819ec3c2244577ab77bdd0b9a5512ba3a7cb61ae8bafb4/userdata/shm DeviceMajor:0 DeviceMinor:473 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-133 DeviceMajor:0 DeviceMinor:133 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8/volumes/kubernetes.io~projected/kube-api-access-jrdsv DeviceMajor:0 DeviceMinor:138 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-392 DeviceMajor:0 DeviceMinor:392 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/eecf6d3a967ce077414ad07bec5a21f5acf63791b8f14de4d8d34c6cfe91444a/userdata/shm DeviceMajor:0 DeviceMinor:468 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a/volumes/kubernetes.io~projected/kube-api-access-pdsh9 DeviceMajor:0 DeviceMinor:472 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/6ead10efe7a643d9b0dd883cc19f8ef852a5658bc79c235b7c7c6b5de2e97811/userdata/shm DeviceMajor:0 DeviceMinor:189 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~secret/cluster-olm-operator-serving-cert DeviceMajor:0 DeviceMinor:263 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-199 DeviceMajor:0 DeviceMinor:199 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-268 DeviceMajor:0 DeviceMinor:268 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-320 DeviceMajor:0 DeviceMinor:320 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-406 DeviceMajor:0 DeviceMinor:406 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-408 DeviceMajor:0 DeviceMinor:408 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-78 DeviceMajor:0 DeviceMinor:78 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-169 DeviceMajor:0 DeviceMinor:169 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-129 DeviceMajor:0 DeviceMinor:129 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-154 DeviceMajor:0 DeviceMinor:154 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/fc75409854b1370368a39cd91bcd5c4ac2aa319c2a8b9aaad1e7abab3fd56a49/userdata/shm DeviceMajor:0 DeviceMinor:179 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-244 DeviceMajor:0 DeviceMinor:244 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:255 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~projected/kube-api-access-2fbss DeviceMajor:0 DeviceMinor:313 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-48 DeviceMajor:0 DeviceMinor:48 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-82 DeviceMajor:0 DeviceMinor:82 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-362 DeviceMajor:0 DeviceMinor:362 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/87b1749af9a101502a0aa9cd2d75f0e1562789f11a001117a5d269d9ed620723/userdata/shm DeviceMajor:0 DeviceMinor:58 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-191 DeviceMajor:0 DeviceMinor:191 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~projected/kube-api-access-jv4zs DeviceMajor:0 DeviceMinor:286 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/8d76404b-6d62-4a61-b6f6-0c8073eba198/volumes/kubernetes.io~projected/kube-api-access-7kb5c DeviceMajor:0 DeviceMinor:293 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~projected/kube-api-access-twh85 DeviceMajor:0 DeviceMinor:299 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/926263c4-ec5b-41cb-9c30-0c88f636035f/volumes/kubernetes.io~projected/kube-api-access-fncz7 DeviceMajor:0 DeviceMinor:318 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-44 DeviceMajor:0 DeviceMinor:44 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/7672f24a6f45e585db5dc690d30bd20c69d755aa3cc9d4346ea6ca7923305116/userdata/shm DeviceMajor:0 DeviceMinor:54 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-381 DeviceMajor:0 DeviceMinor:381 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-394 DeviceMajor:0 DeviceMinor:394 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-390 DeviceMajor:0 DeviceMinor:390 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-470 DeviceMajor:0 DeviceMinor:470 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-489 DeviceMajor:0 DeviceMinor:489 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes.io~projected/kube-api-access-fzwfq DeviceMajor:0 DeviceMinor:266 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~projected/kube-api-access DeviceMajor:0 DeviceMinor:281 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/6a9f011f-36f1-4308-a365-69425c186c7f/volumes/kubernetes.io~projected/kube-api-access-mvnxf DeviceMajor:0 DeviceMinor:137 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:264 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/12d79e8d025da0e5c89728d8b35230c8c9764259d39417d2ff38f00dc9c263f6/userdata/shm DeviceMajor:0 DeviceMinor:349 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-354 DeviceMajor:0 DeviceMinor:354 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:260 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-236 DeviceMajor:0 DeviceMinor:236 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-62 DeviceMajor:0 DeviceMinor:62 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-64 DeviceMajor:0 DeviceMinor:64 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-80 DeviceMajor:0 DeviceMinor:80 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-90 DeviceMajor:0 DeviceMinor:90 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-183 DeviceMajor:0 DeviceMinor:183 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~projected/kube-api-access-wg9zq DeviceMajor:0 DeviceMinor:188 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/283122ba-be1c-4516-bd0f-df41c13c098b/volumes/kubernetes.io~projected/kube-api-access-vhjhk DeviceMajor:0 DeviceMinor:275 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/ccbec542fddb9bd0831349589c4ac65650d79b1cf75a0af75cf8ca2d1a0d89f6/userdata/shm DeviceMajor:0 DeviceMinor:337 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:25257074688 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/611473a6bb698264f5594083fb6219f23f5848bd85b0fdd8dd5932b32fae3c36/userdata/shm DeviceMajor:0 DeviceMinor:50 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-358 DeviceMajor:0 DeviceMinor:358 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-112 DeviceMajor:0 DeviceMinor:112 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-294 DeviceMajor:0 DeviceMinor:294 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:277 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-356 DeviceMajor:0 DeviceMinor:356 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-402 DeviceMajor:0 DeviceMinor:402 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/6607da6ec4214a3fe2d315e70d9756b15e987231bb8310517d849da88c8e7b15/userdata/shm DeviceMajor:0 DeviceMinor:165 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:258 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-483 DeviceMajor:0 DeviceMinor:483 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-181 DeviceMajor:0 DeviceMinor:181 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/0d2211dcab91af9edc388c398c5fc7a90c2e8c65270b4ebc9dead574b96eaefe/userdata/shm DeviceMajor:0 DeviceMinor:332 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-195 DeviceMajor:0 DeviceMinor:195 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-209 DeviceMajor:0 DeviceMinor:209 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-197 DeviceMajor:0 DeviceMinor:197 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~projected/kube-api-access-qn7rj DeviceMajor:0 DeviceMinor:267 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-322 DeviceMajor:0 DeviceMinor:322 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-364 DeviceMajor:0 DeviceMinor:364 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~secret/metrics-tls DeviceMajor:0 DeviceMinor:43 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-193 DeviceMajor:0 DeviceMinor:193 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-387 DeviceMajor:0 DeviceMinor:387 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-377 DeviceMajor:0 DeviceMinor:377 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-120 DeviceMajor:0 DeviceMinor:120 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-167 DeviceMajor:0 DeviceMinor:167 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~projected/kube-api-access DeviceMajor:0 DeviceMinor:288 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/bound-sa-token DeviceMajor:0 DeviceMinor:312 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-60 DeviceMajor:0 DeviceMinor:60 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~projected/kube-api-access-jfz87 DeviceMajor:0 DeviceMinor:164 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-156 DeviceMajor:0 DeviceMinor:156 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~projected/kube-api-access DeviceMajor:0 DeviceMinor:290 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/b44ac27d1b7f531e11c3f4d352d89c4a7679a795ec64417e43c87c5d1e725e90/userdata/shm DeviceMajor:0 DeviceMinor:314 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/e2506cc6607d77e49135fcb375a37bafe2d511f6ed75718677873e40b092ab87/userdata/shm DeviceMajor:0 DeviceMinor:333 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-398 DeviceMajor:0 DeviceMinor:398 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-404 DeviceMajor:0 DeviceMinor:404 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~projected/kube-api-access-cwqkb DeviceMajor:0 DeviceMinor:122 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-151 DeviceMajor:0 DeviceMinor:151 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:262 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-400 DeviceMajor:0 DeviceMinor:400 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/89c8e517113e00942d58576e87acb9fd6a3e8f726e30d612913a3d827c50d10f/userdata/shm DeviceMajor:0 DeviceMinor:485 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-147 DeviceMajor:0 DeviceMinor:147 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-185 DeviceMajor:0 DeviceMinor:185 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/etcd-client DeviceMajor:0 DeviceMinor:261 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/5ab705ed0c0dfea5ee8845410f03f5084071123b3566b80a7ef7e6fcce0fa974/userdata/shm DeviceMajor:0 DeviceMinor:383 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-131 DeviceMajor:0 DeviceMinor:131 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-252 DeviceMajor:0 DeviceMinor:252 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/5a1bdc70-6412-47e0-8330-04d796cc8d55/volumes/kubernetes.io~projected/kube-api-access-5dtfn DeviceMajor:0 DeviceMinor:292 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-104 DeviceMajor:0 DeviceMinor:104 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-135 DeviceMajor:0 DeviceMinor:135 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/7e22b55ec9c49898cca9f69ea15552e8a230cbb6781fbf80f0e4919b22b7b7d8/userdata/shm DeviceMajor:0 DeviceMinor:338 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/4833f667ac4451a150ed477789be72c08b65162d8c7bbf29e44cd656299dbd63/userdata/shm DeviceMajor:0 DeviceMinor:375 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~secret/ovn-node-metrics-cert DeviceMajor:0 DeviceMinor:171 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-228 DeviceMajor:0 DeviceMinor:228 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none} 252:16:{Name:vdb Major:252 Minor:16 Size:21474836480 Scheduler:none} 252:32:{Name:vdc Major:252 Minor:32 Size:21474836480 Scheduler:none} 252:48:{Name:vdd Major:252 Minor:48 Size:21474836480 Scheduler:none} 252:64:{Name:vde Major:252 Minor:64 Size:21474836480 Scheduler:none}] NetworkDevices:[{Name:0d2211dcab91af9 MacAddress:ce:f0:13:83:60:f8 Speed:10000 Mtu:8900} {Name:12d79e8d025da0e MacAddress:b2:de:cd:56:04:ba Speed:10000 Mtu:8900} {Name:4833f667ac4451a MacAddress:0e:d6:8b:f6:6f:9f Speed:10000 Mtu:8900} {Name:5ab705ed0c0dfea MacAddress:66:13:58:48:4d:0c Speed:10000 Mtu:8900} {Name:5b905a3a36c591b MacAddress:4a:f7:24:e5:b8:21 Speed:10000 Mtu:8900} {Name:70e7f53d39e81bf MacAddress:fe:c7:bc:67:b8:7b Speed:10000 Mtu:8900} {Name:7e22b55ec9c4989 MacAddress:f6:bd:3e:84:a6:be Speed:10000 Mtu:8900} {Name:8114256db030142 MacAddress:56:1b:41:13:2d:bf Speed:10000 Mtu:8900} {Name:89c8e517113e009 MacAddress:b2:60:01:03:12:96 Speed:10000 Mtu:8900} {Name:b7452830de99c04 MacAddress:c2:1d:24:b1:50:ac Speed:10000 Mtu:8900} {Name:bdfbd9daf37e036 MacAddress:de:e9:9b:b0:9b:e1 Speed:10000 Mtu:8900} {Name:br-ex MacAddress:fa:16:9e:81:f6:10 Speed:0 Mtu:9000} {Name:br-int MacAddress:d6:69:d9:06:a4:58 Speed:0 Mtu:8900} {Name:ccbec542fddb9bd MacAddress:4a:ef:f1:f6:11:27 Speed:10000 Mtu:8900} {Name:e2506cc6607d77e MacAddress:6a:bb:3f:ad:f7:51 Speed:10000 Mtu:8900} {Name:eecf6d3a967ce07 MacAddress:fa:0a:6b:a5:56:79 Speed:10000 Mtu:8900} {Name:eth0 MacAddress:fa:16:9e:81:f6:10 Speed:-1 Mtu:9000} {Name:eth1 MacAddress:fa:16:3e:5e:6c:eb Speed:-1 Mtu:9000} {Name:eth2 MacAddress:fa:16:3e:5e:ad:70 Speed:-1 Mtu:9000} {Name:fedfe3c8bf160c8 MacAddress:8a:81:32:e2:c5:80 Speed:10000 Mtu:8900} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:80:00:02 Speed:0 Mtu:8900} {Name:ovs-system MacAddress:6e:04:ec:85:c4:7f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:50514153472 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[12] Caches:[{Id:12 Size:32768 Type:Data Level:1} {Id:12 Size:32768 Type:Instruction Level:1} {Id:12 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:12 Size:16777216 Type:Unified Level:3}] SocketID:12 BookID: DrawerID:} {Id:0 Threads:[13] Caches:[{Id:13 Size:32768 Type:Data Level:1} {Id:13 Size:32768 Type:Instruction Level:1} {Id:13 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:13 Size:16777216 Type:Unified Level:3}] SocketID:13 BookID: DrawerID:} {Id:0 Threads:[14] Caches:[{Id:14 Size:32768 Type:Data Level:1} {Id:14 Size:32768 Type:Instruction Level:1} {Id:14 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:14 Size:16777216 Type:Unified Level:3}] SocketID:14 BookID: DrawerID:} {Id:0 Threads:[15] Caches:[{Id:15 Size:32768 Type:Data Level:1} {Id:15 Size:32768 Type:Instruction Level:1} {Id:15 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:15 Size:16777216 Type:Unified Level:3}] SocketID:15 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 05 10:37:39.522338 master-0 kubenswrapper[10444]: I1205 10:37:39.522312 10444 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 05 10:37:39.522881 master-0 kubenswrapper[10444]: I1205 10:37:39.522533 10444 manager.go:233] Version: {KernelVersion:5.14.0-427.100.1.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202511170715-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 05 10:37:39.522881 master-0 kubenswrapper[10444]: I1205 10:37:39.522844 10444 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 05 10:37:39.523021 master-0 kubenswrapper[10444]: I1205 10:37:39.522978 10444 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 05 10:37:39.523183 master-0 kubenswrapper[10444]: I1205 10:37:39.523012 10444 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"master-0","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 05 10:37:39.523247 master-0 kubenswrapper[10444]: I1205 10:37:39.523196 10444 topology_manager.go:138] "Creating topology manager with none policy" Dec 05 10:37:39.523247 master-0 kubenswrapper[10444]: I1205 10:37:39.523206 10444 container_manager_linux.go:303] "Creating device plugin manager" Dec 05 10:37:39.523247 master-0 kubenswrapper[10444]: I1205 10:37:39.523214 10444 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 10:37:39.523247 master-0 kubenswrapper[10444]: I1205 10:37:39.523234 10444 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 10:37:39.523447 master-0 kubenswrapper[10444]: I1205 10:37:39.523388 10444 state_mem.go:36] "Initialized new in-memory state store" Dec 05 10:37:39.523525 master-0 kubenswrapper[10444]: I1205 10:37:39.523500 10444 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 05 10:37:39.523567 master-0 kubenswrapper[10444]: I1205 10:37:39.523558 10444 kubelet.go:418] "Attempting to sync node with API server" Dec 05 10:37:39.523604 master-0 kubenswrapper[10444]: I1205 10:37:39.523569 10444 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 05 10:37:39.523604 master-0 kubenswrapper[10444]: I1205 10:37:39.523584 10444 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 05 10:37:39.523604 master-0 kubenswrapper[10444]: I1205 10:37:39.523595 10444 kubelet.go:324] "Adding apiserver pod source" Dec 05 10:37:39.523714 master-0 kubenswrapper[10444]: I1205 10:37:39.523619 10444 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 05 10:37:39.524950 master-0 kubenswrapper[10444]: I1205 10:37:39.524911 10444 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.13-2.rhaos4.18.git15789b8.el9" apiVersion="v1" Dec 05 10:37:39.525223 master-0 kubenswrapper[10444]: I1205 10:37:39.525188 10444 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 05 10:37:39.525574 master-0 kubenswrapper[10444]: I1205 10:37:39.525550 10444 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 05 10:37:39.525836 master-0 kubenswrapper[10444]: I1205 10:37:39.525804 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 05 10:37:39.525836 master-0 kubenswrapper[10444]: I1205 10:37:39.525833 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 05 10:37:39.525939 master-0 kubenswrapper[10444]: I1205 10:37:39.525843 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 05 10:37:39.525939 master-0 kubenswrapper[10444]: I1205 10:37:39.525852 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 05 10:37:39.525939 master-0 kubenswrapper[10444]: I1205 10:37:39.525861 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 05 10:37:39.525939 master-0 kubenswrapper[10444]: I1205 10:37:39.525870 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 05 10:37:39.525939 master-0 kubenswrapper[10444]: I1205 10:37:39.525879 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 05 10:37:39.525939 master-0 kubenswrapper[10444]: I1205 10:37:39.525887 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 05 10:37:39.525939 master-0 kubenswrapper[10444]: I1205 10:37:39.525898 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 05 10:37:39.525939 master-0 kubenswrapper[10444]: I1205 10:37:39.525908 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 05 10:37:39.525939 master-0 kubenswrapper[10444]: I1205 10:37:39.525948 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 05 10:37:39.526242 master-0 kubenswrapper[10444]: I1205 10:37:39.525967 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 05 10:37:39.526242 master-0 kubenswrapper[10444]: I1205 10:37:39.526001 10444 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 05 10:37:39.526661 master-0 kubenswrapper[10444]: I1205 10:37:39.526638 10444 server.go:1280] "Started kubelet" Dec 05 10:37:39.526808 master-0 kubenswrapper[10444]: I1205 10:37:39.526733 10444 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 05 10:37:39.527075 master-0 kubenswrapper[10444]: I1205 10:37:39.526960 10444 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 05 10:37:39.527161 master-0 kubenswrapper[10444]: I1205 10:37:39.527119 10444 server_v1.go:47] "podresources" method="list" useActivePods=true Dec 05 10:37:39.527854 master-0 kubenswrapper[10444]: I1205 10:37:39.527806 10444 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 05 10:37:39.528154 master-0 kubenswrapper[10444]: I1205 10:37:39.528126 10444 server.go:449] "Adding debug handlers to kubelet server" Dec 05 10:37:39.529571 master-0 systemd[1]: Started Kubernetes Kubelet. Dec 05 10:37:39.546090 master-0 kubenswrapper[10444]: I1205 10:37:39.545924 10444 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 05 10:37:39.546090 master-0 kubenswrapper[10444]: I1205 10:37:39.545973 10444 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 05 10:37:39.546090 master-0 kubenswrapper[10444]: I1205 10:37:39.546006 10444 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2025-12-06 10:27:25 +0000 UTC, rotation deadline is 2025-12-06 03:49:39.508039355 +0000 UTC Dec 05 10:37:39.546090 master-0 kubenswrapper[10444]: I1205 10:37:39.546047 10444 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 17h11m59.961995041s for next certificate rotation Dec 05 10:37:39.546090 master-0 kubenswrapper[10444]: I1205 10:37:39.546096 10444 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 05 10:37:39.546502 master-0 kubenswrapper[10444]: I1205 10:37:39.546113 10444 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 05 10:37:39.546502 master-0 kubenswrapper[10444]: I1205 10:37:39.546247 10444 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Dec 05 10:37:39.546502 master-0 kubenswrapper[10444]: E1205 10:37:39.546111 10444 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"master-0\" not found" Dec 05 10:37:39.551149 master-0 kubenswrapper[10444]: I1205 10:37:39.551024 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" volumeName="kubernetes.io/projected/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-kube-api-access-2fbss" seLinuxMountContext="" Dec 05 10:37:39.551149 master-0 kubenswrapper[10444]: I1205 10:37:39.551134 10444 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 05 10:37:39.551149 master-0 kubenswrapper[10444]: I1205 10:37:39.551158 10444 factory.go:55] Registering systemd factory Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551176 10444 factory.go:221] Registration of the systemd container factory successfully Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551138 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1a0235af-2cf2-4ad4-b419-764fb56a0107" volumeName="kubernetes.io/projected/1a0235af-2cf2-4ad4-b419-764fb56a0107-kube-api-access-cwqkb" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551260 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1e69ce9e-4e6f-4015-9ba6-5a7942570190" volumeName="kubernetes.io/projected/1e69ce9e-4e6f-4015-9ba6-5a7942570190-kube-api-access-7wfsv" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551295 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="926263c4-ec5b-41cb-9c30-0c88f636035f" volumeName="kubernetes.io/projected/926263c4-ec5b-41cb-9c30-0c88f636035f-kube-api-access-fncz7" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551309 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c22d947f-a5b6-4f24-b142-dd201c46293b" volumeName="kubernetes.io/projected/c22d947f-a5b6-4f24-b142-dd201c46293b-kube-api-access-98qt8" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551322 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d1c3b7dd-f25e-4983-8a94-084f863fd5b9" volumeName="kubernetes.io/configmap/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-config" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551334 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="11f563d5-89bb-433c-956a-6d5d2492e8f1" volumeName="kubernetes.io/configmap/11f563d5-89bb-433c-956a-6d5d2492e8f1-config" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551349 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4825316a-ea9f-4d3d-838b-fa809a6e49c7" volumeName="kubernetes.io/configmap/4825316a-ea9f-4d3d-838b-fa809a6e49c7-config" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551368 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9f011f-36f1-4308-a365-69425c186c7f" volumeName="kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-cni-binary-copy" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551381 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6f76d12f-5406-47e2-8337-2f50e35376d6" volumeName="kubernetes.io/configmap/6f76d12f-5406-47e2-8337-2f50e35376d6-config" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551393 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d1c3b7dd-f25e-4983-8a94-084f863fd5b9" volumeName="kubernetes.io/projected/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-kube-api-access-fkmvj" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551405 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e" volumeName="kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-env-overrides" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551437 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7b29f89-e42d-4e53-ad14-05efdce933f0" volumeName="kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-kube-api-access-khfxz" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551455 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22676fac-b770-4937-9bee-7478bd1babb7" volumeName="kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-kube-api-access-vkqz7" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551467 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="495ba1ea-f844-43ec-8be7-47e738f5428a" volumeName="kubernetes.io/projected/495ba1ea-f844-43ec-8be7-47e738f5428a-kube-api-access-tjgc4" seLinuxMountContext="" Dec 05 10:37:39.551466 master-0 kubenswrapper[10444]: I1205 10:37:39.551480 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fd58232c-a81a-4aee-8b2c-5ffcdded2e23" volumeName="kubernetes.io/projected/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-kube-api-access-h2vd4" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551494 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b760849c-8d83-47da-8677-68445c143bef" volumeName="kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-env-overrides" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551508 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-serving-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551499 10444 factory.go:153] Registering CRI-O factory Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551541 10444 factory.go:221] Registration of the crio container factory successfully Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551567 10444 factory.go:103] Registering Raw factory Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551580 10444 manager.go:1196] Started watching for new ooms in manager Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551521 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="720a7467-ce93-4d48-82ec-9ad0922d99c2" volumeName="kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551864 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8d76404b-6d62-4a61-b6f6-0c8073eba198" volumeName="kubernetes.io/projected/8d76404b-6d62-4a61-b6f6-0c8073eba198-kube-api-access-7kb5c" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551897 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d1c3b7dd-f25e-4983-8a94-084f863fd5b9" volumeName="kubernetes.io/secret/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-serving-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551918 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" volumeName="kubernetes.io/secret/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-serving-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551937 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e27c0798-ec1c-43cd-b81b-f77f2f11ad0f" volumeName="kubernetes.io/projected/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f-kube-api-access-ml74w" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551955 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1a0235af-2cf2-4ad4-b419-764fb56a0107" volumeName="kubernetes.io/secret/1a0235af-2cf2-4ad4-b419-764fb56a0107-metrics-tls" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551971 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082" volumeName="kubernetes.io/projected/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-kube-api-access-jv4zs" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551987 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4825316a-ea9f-4d3d-838b-fa809a6e49c7" volumeName="kubernetes.io/projected/4825316a-ea9f-4d3d-838b-fa809a6e49c7-kube-api-access" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552009 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="82ef99d4-41b7-4859-a5a1-4e31ce614a2a" volumeName="kubernetes.io/projected/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-kube-api-access" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552029 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b760849c-8d83-47da-8677-68445c143bef" volumeName="kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-ovnkube-config" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552047 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" volumeName="kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-trusted-ca-bundle" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552062 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-ca" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552078 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" volumeName="kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-whereabouts-configmap" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552094 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="444f8808-e454-4015-9e20-429e715a08c7" volumeName="kubernetes.io/projected/444f8808-e454-4015-9e20-429e715a08c7-kube-api-access" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552111 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="495ba1ea-f844-43ec-8be7-47e738f5428a" volumeName="kubernetes.io/secret/495ba1ea-f844-43ec-8be7-47e738f5428a-ovn-node-metrics-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552141 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c22d947f-a5b6-4f24-b142-dd201c46293b" volumeName="kubernetes.io/empty-dir/c22d947f-a5b6-4f24-b142-dd201c46293b-available-featuregates" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552157 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e" volumeName="kubernetes.io/projected/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-kube-api-access-wg9zq" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552173 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-service-ca" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552193 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" volumeName="kubernetes.io/projected/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-kube-api-access-ssssf" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552209 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4825316a-ea9f-4d3d-838b-fa809a6e49c7" volumeName="kubernetes.io/secret/4825316a-ea9f-4d3d-838b-fa809a6e49c7-serving-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552225 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6f76d12f-5406-47e2-8337-2f50e35376d6" volumeName="kubernetes.io/projected/6f76d12f-5406-47e2-8337-2f50e35376d6-kube-api-access-fzwfq" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552239 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="82ef99d4-41b7-4859-a5a1-4e31ce614a2a" volumeName="kubernetes.io/configmap/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-service-ca" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552254 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="11f563d5-89bb-433c-956a-6d5d2492e8f1" volumeName="kubernetes.io/projected/11f563d5-89bb-433c-956a-6d5d2492e8f1-kube-api-access" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552270 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3c4f767d-65f9-434b-8ddd-ceb0b91ab99a" volumeName="kubernetes.io/projected/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a-kube-api-access-pdsh9" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552284 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49051e6e-5a2f-45c8-bad0-374514a91c07" volumeName="kubernetes.io/projected/49051e6e-5a2f-45c8-bad0-374514a91c07-kube-api-access-6vp9b" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552297 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="495ba1ea-f844-43ec-8be7-47e738f5428a" volumeName="kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-script-lib" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552314 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a722cda9-29a0-4b7f-8e1d-9a8950ed765a" volumeName="kubernetes.io/configmap/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-telemetry-config" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552355 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b760849c-8d83-47da-8677-68445c143bef" volumeName="kubernetes.io/projected/b760849c-8d83-47da-8677-68445c143bef-kube-api-access-jfz87" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552372 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e" volumeName="kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-ovnkube-identity-cm" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552386 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="eb290494-a456-4f0e-9afc-f20abab1a1bf" volumeName="kubernetes.io/projected/eb290494-a456-4f0e-9afc-f20abab1a1bf-kube-api-access-qn7rj" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552411 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="283122ba-be1c-4516-bd0f-df41c13c098b" volumeName="kubernetes.io/projected/283122ba-be1c-4516-bd0f-df41c13c098b-kube-api-access-vhjhk" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552428 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" volumeName="kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-binary-copy" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552467 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8c649a16-c187-412e-b5da-62a00bee38ab" volumeName="kubernetes.io/projected/8c649a16-c187-412e-b5da-62a00bee38ab-kube-api-access-d8pv2" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552483 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7b29f89-e42d-4e53-ad14-05efdce933f0" volumeName="kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-bound-sa-token" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552573 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="11f563d5-89bb-433c-956a-6d5d2492e8f1" volumeName="kubernetes.io/secret/11f563d5-89bb-433c-956a-6d5d2492e8f1-serving-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552600 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="495ba1ea-f844-43ec-8be7-47e738f5428a" volumeName="kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-env-overrides" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552615 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a722cda9-29a0-4b7f-8e1d-9a8950ed765a" volumeName="kubernetes.io/projected/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-kube-api-access-2lwgq" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552631 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/projected/f7a08359-0379-4364-8b0c-ddb58ff605f4-kube-api-access-twh85" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552646 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="283122ba-be1c-4516-bd0f-df41c13c098b" volumeName="kubernetes.io/configmap/283122ba-be1c-4516-bd0f-df41c13c098b-iptables-alerter-script" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552661 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6f76d12f-5406-47e2-8337-2f50e35376d6" volumeName="kubernetes.io/secret/6f76d12f-5406-47e2-8337-2f50e35376d6-serving-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552677 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="444f8808-e454-4015-9e20-429e715a08c7" volumeName="kubernetes.io/configmap/444f8808-e454-4015-9e20-429e715a08c7-config" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552691 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49051e6e-5a2f-45c8-bad0-374514a91c07" volumeName="kubernetes.io/empty-dir/49051e6e-5a2f-45c8-bad0-374514a91c07-operand-assets" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552707 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49051e6e-5a2f-45c8-bad0-374514a91c07" volumeName="kubernetes.io/secret/49051e6e-5a2f-45c8-bad0-374514a91c07-cluster-olm-operator-serving-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552724 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5a1bdc70-6412-47e0-8330-04d796cc8d55" volumeName="kubernetes.io/projected/5a1bdc70-6412-47e0-8330-04d796cc8d55-kube-api-access-5dtfn" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552741 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9f011f-36f1-4308-a365-69425c186c7f" volumeName="kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-multus-daemon-config" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552757 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87909f47-f2d7-46f8-a1c8-27336cdcce5d" volumeName="kubernetes.io/projected/87909f47-f2d7-46f8-a1c8-27336cdcce5d-kube-api-access-rjrgm" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552773 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22676fac-b770-4937-9bee-7478bd1babb7" volumeName="kubernetes.io/configmap/22676fac-b770-4937-9bee-7478bd1babb7-trusted-ca" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552791 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" volumeName="kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-sysctl-allowlist" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552807 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fd58232c-a81a-4aee-8b2c-5ffcdded2e23" volumeName="kubernetes.io/configmap/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-config" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552824 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c22d947f-a5b6-4f24-b142-dd201c46293b" volumeName="kubernetes.io/secret/c22d947f-a5b6-4f24-b142-dd201c46293b-serving-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552843 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" volumeName="kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-config" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552859 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e" volumeName="kubernetes.io/secret/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-webhook-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552874 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-config" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552890 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fd58232c-a81a-4aee-8b2c-5ffcdded2e23" volumeName="kubernetes.io/secret/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-serving-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552904 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082" volumeName="kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-profile-collector-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552919 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="495ba1ea-f844-43ec-8be7-47e738f5428a" volumeName="kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-config" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552935 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9f011f-36f1-4308-a365-69425c186c7f" volumeName="kubernetes.io/projected/6a9f011f-36f1-4308-a365-69425c186c7f-kube-api-access-mvnxf" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552952 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b760849c-8d83-47da-8677-68445c143bef" volumeName="kubernetes.io/secret/b760849c-8d83-47da-8677-68445c143bef-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552967 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" volumeName="kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-service-ca-bundle" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.552984 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="eb290494-a456-4f0e-9afc-f20abab1a1bf" volumeName="kubernetes.io/configmap/eb290494-a456-4f0e-9afc-f20abab1a1bf-trusted-ca" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.553000 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-client" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.553015 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7b29f89-e42d-4e53-ad14-05efdce933f0" volumeName="kubernetes.io/configmap/f7b29f89-e42d-4e53-ad14-05efdce933f0-trusted-ca" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.553030 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22676fac-b770-4937-9bee-7478bd1babb7" volumeName="kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-bound-sa-token" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.553045 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" volumeName="kubernetes.io/projected/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-kube-api-access-jrdsv" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.553061 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="926263c4-ec5b-41cb-9c30-0c88f636035f" volumeName="kubernetes.io/configmap/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-trusted-ca" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.553079 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1e69ce9e-4e6f-4015-9ba6-5a7942570190" volumeName="kubernetes.io/configmap/1e69ce9e-4e6f-4015-9ba6-5a7942570190-config" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.553098 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1e69ce9e-4e6f-4015-9ba6-5a7942570190" volumeName="kubernetes.io/secret/1e69ce9e-4e6f-4015-9ba6-5a7942570190-serving-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.553114 10444 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="444f8808-e454-4015-9e20-429e715a08c7" volumeName="kubernetes.io/secret/444f8808-e454-4015-9e20-429e715a08c7-serving-cert" seLinuxMountContext="" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.553130 10444 reconstruct.go:97] "Volume reconstruction finished" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.553142 10444 reconciler.go:26] "Reconciler: start to sync state" Dec 05 10:37:39.553171 master-0 kubenswrapper[10444]: I1205 10:37:39.551962 10444 manager.go:319] Starting recovery of all containers Dec 05 10:37:39.557348 master-0 kubenswrapper[10444]: I1205 10:37:39.556130 10444 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 10:37:39.557348 master-0 kubenswrapper[10444]: I1205 10:37:39.556671 10444 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 10:37:39.557348 master-0 kubenswrapper[10444]: I1205 10:37:39.556859 10444 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 10:37:39.562412 master-0 kubenswrapper[10444]: I1205 10:37:39.562325 10444 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 05 10:37:39.566319 master-0 kubenswrapper[10444]: E1205 10:37:39.566103 10444 kubelet.go:1495] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Dec 05 10:37:39.590179 master-0 kubenswrapper[10444]: I1205 10:37:39.589728 10444 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 05 10:37:39.592500 master-0 kubenswrapper[10444]: I1205 10:37:39.592467 10444 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 05 10:37:39.592617 master-0 kubenswrapper[10444]: I1205 10:37:39.592507 10444 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 05 10:37:39.592617 master-0 kubenswrapper[10444]: I1205 10:37:39.592573 10444 kubelet.go:2335] "Starting kubelet main sync loop" Dec 05 10:37:39.592675 master-0 kubenswrapper[10444]: E1205 10:37:39.592616 10444 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 05 10:37:39.594687 master-0 kubenswrapper[10444]: I1205 10:37:39.594665 10444 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 10:37:39.602400 master-0 kubenswrapper[10444]: I1205 10:37:39.602360 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_kube-rbac-proxy-crio-master-0_3169f44496ed8a28c6d6a15511ab0eec/kube-rbac-proxy-crio/2.log" Dec 05 10:37:39.602780 master-0 kubenswrapper[10444]: I1205 10:37:39.602719 10444 generic.go:334] "Generic (PLEG): container finished" podID="3169f44496ed8a28c6d6a15511ab0eec" containerID="c4c2663f7fd33f4762e4466ae563543a600762863049b80576e8bb073c24bbd0" exitCode=1 Dec 05 10:37:39.602780 master-0 kubenswrapper[10444]: I1205 10:37:39.602769 10444 generic.go:334] "Generic (PLEG): container finished" podID="3169f44496ed8a28c6d6a15511ab0eec" containerID="2a995b4c36846fd814a8f2ca3bd01eb60eaabd9a004a171a036b25b030f63eff" exitCode=0 Dec 05 10:37:39.611328 master-0 kubenswrapper[10444]: I1205 10:37:39.611266 10444 generic.go:334] "Generic (PLEG): container finished" podID="f1d74f88-1419-431e-80da-26db419f050e" containerID="afde21a1bd7da8808580dd974998cee92cb64026aa792ec094def214f272bc2f" exitCode=0 Dec 05 10:37:39.614654 master-0 kubenswrapper[10444]: I1205 10:37:39.614604 10444 generic.go:334] "Generic (PLEG): container finished" podID="8b47694fcc32464ab24d09c23d6efb57" containerID="6c524eb37d9eb519540115909748ea1e0272427fa38212f1b6de731f3f942edd" exitCode=1 Dec 05 10:37:39.616328 master-0 kubenswrapper[10444]: I1205 10:37:39.616304 10444 generic.go:334] "Generic (PLEG): container finished" podID="c22d947f-a5b6-4f24-b142-dd201c46293b" containerID="d0d4427b1c533edfc1d254b4b4202c620d108a1b343169a0fecb4359d771c631" exitCode=0 Dec 05 10:37:39.622267 master-0 kubenswrapper[10444]: I1205 10:37:39.622217 10444 generic.go:334] "Generic (PLEG): container finished" podID="49051e6e-5a2f-45c8-bad0-374514a91c07" containerID="be912c90bc4a342d66fa17918d807831fa70aa4a080f3751a563cdc9efce0a37" exitCode=0 Dec 05 10:37:39.628486 master-0 kubenswrapper[10444]: I1205 10:37:39.628414 10444 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="12675055319a88464b4e8137b9c8f20eb9ba5b578bde42cad92050f544a0c6be" exitCode=0 Dec 05 10:37:39.628486 master-0 kubenswrapper[10444]: I1205 10:37:39.628461 10444 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="ee9e5b75d34cda6a745ee4f00c4f147db05bd9943033e03d413d4fa85d63bb34" exitCode=0 Dec 05 10:37:39.628486 master-0 kubenswrapper[10444]: I1205 10:37:39.628472 10444 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="3acdc4bc20dace9999e3e38bd13059d0977882b304881b0e3ccd402e31f993c2" exitCode=0 Dec 05 10:37:39.628486 master-0 kubenswrapper[10444]: I1205 10:37:39.628481 10444 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="f4f209477f798940668277c240fd1326e8fd7d7c2b8eedd2830e59be706883e8" exitCode=0 Dec 05 10:37:39.628486 master-0 kubenswrapper[10444]: I1205 10:37:39.628490 10444 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="215c293fe494f6f41ad17e025bbeffa977136f630518896e92b337f0d84c340a" exitCode=0 Dec 05 10:37:39.629005 master-0 kubenswrapper[10444]: I1205 10:37:39.628499 10444 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="10b806d181204c81aa236ce7e2f93aa28932b9c9f02c11fefdb2f0d1d7bdeac9" exitCode=0 Dec 05 10:37:39.630821 master-0 kubenswrapper[10444]: I1205 10:37:39.630785 10444 generic.go:334] "Generic (PLEG): container finished" podID="3af980e3-b7b6-4a5b-96c4-53cbbe43c268" containerID="53136dd9eb8d1d61f6ae883e0b488e0fb1c3b2414f554f6f62aa4dc092f5012c" exitCode=0 Dec 05 10:37:39.644301 master-0 kubenswrapper[10444]: I1205 10:37:39.644205 10444 generic.go:334] "Generic (PLEG): container finished" podID="495ba1ea-f844-43ec-8be7-47e738f5428a" containerID="f02b670dd80cfe6eb6e6f824971df1d4d23971943a981cfcb563de519950bb3e" exitCode=0 Dec 05 10:37:39.653286 master-0 kubenswrapper[10444]: I1205 10:37:39.653199 10444 generic.go:334] "Generic (PLEG): container finished" podID="d75143d9bc4a2dc15781dc51ccff632a" containerID="738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948" exitCode=0 Dec 05 10:37:39.692737 master-0 kubenswrapper[10444]: E1205 10:37:39.692693 10444 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 05 10:37:39.725935 master-0 kubenswrapper[10444]: I1205 10:37:39.725896 10444 manager.go:324] Recovery completed Dec 05 10:37:39.763720 master-0 kubenswrapper[10444]: I1205 10:37:39.763037 10444 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 05 10:37:39.763720 master-0 kubenswrapper[10444]: I1205 10:37:39.763061 10444 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 05 10:37:39.763720 master-0 kubenswrapper[10444]: I1205 10:37:39.763093 10444 state_mem.go:36] "Initialized new in-memory state store" Dec 05 10:37:39.763720 master-0 kubenswrapper[10444]: I1205 10:37:39.763285 10444 state_mem.go:88] "Updated default CPUSet" cpuSet="" Dec 05 10:37:39.763720 master-0 kubenswrapper[10444]: I1205 10:37:39.763297 10444 state_mem.go:96] "Updated CPUSet assignments" assignments={} Dec 05 10:37:39.763720 master-0 kubenswrapper[10444]: I1205 10:37:39.763316 10444 state_checkpoint.go:136] "State checkpoint: restored state from checkpoint" Dec 05 10:37:39.763720 master-0 kubenswrapper[10444]: I1205 10:37:39.763322 10444 state_checkpoint.go:137] "State checkpoint: defaultCPUSet" defaultCpuSet="" Dec 05 10:37:39.763720 master-0 kubenswrapper[10444]: I1205 10:37:39.763329 10444 policy_none.go:49] "None policy: Start" Dec 05 10:37:39.768683 master-0 kubenswrapper[10444]: I1205 10:37:39.768651 10444 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 05 10:37:39.768785 master-0 kubenswrapper[10444]: I1205 10:37:39.768718 10444 state_mem.go:35] "Initializing new in-memory state store" Dec 05 10:37:39.769047 master-0 kubenswrapper[10444]: I1205 10:37:39.769025 10444 state_mem.go:75] "Updated machine memory state" Dec 05 10:37:39.769047 master-0 kubenswrapper[10444]: I1205 10:37:39.769045 10444 state_checkpoint.go:82] "State checkpoint: restored state from checkpoint" Dec 05 10:37:39.790779 master-0 kubenswrapper[10444]: I1205 10:37:39.790737 10444 manager.go:334] "Starting Device Plugin manager" Dec 05 10:37:39.790779 master-0 kubenswrapper[10444]: I1205 10:37:39.790782 10444 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 05 10:37:39.790779 master-0 kubenswrapper[10444]: I1205 10:37:39.790793 10444 server.go:79] "Starting device plugin registration server" Dec 05 10:37:39.791205 master-0 kubenswrapper[10444]: I1205 10:37:39.791186 10444 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 05 10:37:39.791278 master-0 kubenswrapper[10444]: I1205 10:37:39.791206 10444 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 05 10:37:39.791670 master-0 kubenswrapper[10444]: I1205 10:37:39.791636 10444 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 05 10:37:39.791726 master-0 kubenswrapper[10444]: I1205 10:37:39.791710 10444 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 05 10:37:39.791726 master-0 kubenswrapper[10444]: I1205 10:37:39.791717 10444 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 05 10:37:39.892238 master-0 kubenswrapper[10444]: I1205 10:37:39.892106 10444 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:37:39.893358 master-0 kubenswrapper[10444]: I1205 10:37:39.893266 10444 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-master-0-master-0","openshift-kube-apiserver/bootstrap-kube-apiserver-master-0","kube-system/bootstrap-kube-controller-manager-master-0","kube-system/bootstrap-kube-scheduler-master-0","openshift-machine-config-operator/kube-rbac-proxy-crio-master-0"] Dec 05 10:37:39.893788 master-0 kubenswrapper[10444]: I1205 10:37:39.893714 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerDied","Data":"c4c2663f7fd33f4762e4466ae563543a600762863049b80576e8bb073c24bbd0"} Dec 05 10:37:39.893826 master-0 kubenswrapper[10444]: I1205 10:37:39.893789 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerDied","Data":"2a995b4c36846fd814a8f2ca3bd01eb60eaabd9a004a171a036b25b030f63eff"} Dec 05 10:37:39.893826 master-0 kubenswrapper[10444]: I1205 10:37:39.893806 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerStarted","Data":"d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7"} Dec 05 10:37:39.893877 master-0 kubenswrapper[10444]: I1205 10:37:39.893832 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e72f14b85f51789b3236420aa213917fe54cddf2ceedaab3f6bad48e1ce12513" Dec 05 10:37:39.893877 master-0 kubenswrapper[10444]: I1205 10:37:39.893848 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerStarted","Data":"303d7cff916f69b65cb8e8b72a47e944bb182f86fc3e9bfde505fa8e541eb3b6"} Dec 05 10:37:39.893877 master-0 kubenswrapper[10444]: I1205 10:37:39.893861 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerStarted","Data":"80aee350277cce308d84133aa5eefc6be3ed0712557f82d0d22b78dac448f0af"} Dec 05 10:37:39.893877 master-0 kubenswrapper[10444]: I1205 10:37:39.893872 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerDied","Data":"6c524eb37d9eb519540115909748ea1e0272427fa38212f1b6de731f3f942edd"} Dec 05 10:37:39.893976 master-0 kubenswrapper[10444]: I1205 10:37:39.893897 10444 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:37:39.893976 master-0 kubenswrapper[10444]: I1205 10:37:39.893933 10444 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:37:39.893976 master-0 kubenswrapper[10444]: I1205 10:37:39.893942 10444 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:37:39.894051 master-0 kubenswrapper[10444]: I1205 10:37:39.893995 10444 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:37:39.894111 master-0 kubenswrapper[10444]: I1205 10:37:39.894029 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerStarted","Data":"7672f24a6f45e585db5dc690d30bd20c69d755aa3cc9d4346ea6ca7923305116"} Dec 05 10:37:39.894148 master-0 kubenswrapper[10444]: I1205 10:37:39.894123 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2323b8713dbbb688b1af973566334ca0b77411424d2f12acb62f9e5354b6cd41" Dec 05 10:37:39.894200 master-0 kubenswrapper[10444]: I1205 10:37:39.894170 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" event={"ID":"d75143d9bc4a2dc15781dc51ccff632a","Type":"ContainerStarted","Data":"2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef"} Dec 05 10:37:39.894200 master-0 kubenswrapper[10444]: I1205 10:37:39.894195 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" event={"ID":"d75143d9bc4a2dc15781dc51ccff632a","Type":"ContainerStarted","Data":"10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298"} Dec 05 10:37:39.894259 master-0 kubenswrapper[10444]: I1205 10:37:39.894235 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" event={"ID":"d75143d9bc4a2dc15781dc51ccff632a","Type":"ContainerDied","Data":"738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948"} Dec 05 10:37:39.894288 master-0 kubenswrapper[10444]: I1205 10:37:39.894256 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" event={"ID":"d75143d9bc4a2dc15781dc51ccff632a","Type":"ContainerStarted","Data":"09d6c752fa5b21d6de28695ca1e30252206036602e7b7cad7c82ecf0c820105d"} Dec 05 10:37:39.894325 master-0 kubenswrapper[10444]: I1205 10:37:39.894309 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-scheduler-master-0" event={"ID":"5e09e2af7200e6f9be469dbfd9bb1127","Type":"ContainerStarted","Data":"f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782"} Dec 05 10:37:39.894353 master-0 kubenswrapper[10444]: I1205 10:37:39.894329 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-scheduler-master-0" event={"ID":"5e09e2af7200e6f9be469dbfd9bb1127","Type":"ContainerStarted","Data":"611473a6bb698264f5594083fb6219f23f5848bd85b0fdd8dd5932b32fae3c36"} Dec 05 10:37:39.894353 master-0 kubenswrapper[10444]: I1205 10:37:39.894349 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0-master-0" event={"ID":"cc0396a9a2689b3e8c132c12640cbe83","Type":"ContainerStarted","Data":"99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c"} Dec 05 10:37:39.894402 master-0 kubenswrapper[10444]: I1205 10:37:39.894360 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0-master-0" event={"ID":"cc0396a9a2689b3e8c132c12640cbe83","Type":"ContainerStarted","Data":"619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b"} Dec 05 10:37:39.894402 master-0 kubenswrapper[10444]: I1205 10:37:39.894369 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0-master-0" event={"ID":"cc0396a9a2689b3e8c132c12640cbe83","Type":"ContainerStarted","Data":"87b1749af9a101502a0aa9cd2d75f0e1562789f11a001117a5d269d9ed620723"} Dec 05 10:37:39.905389 master-0 kubenswrapper[10444]: I1205 10:37:39.905364 10444 kubelet_node_status.go:115] "Node was previously registered" node="master-0" Dec 05 10:37:39.905616 master-0 kubenswrapper[10444]: I1205 10:37:39.905603 10444 kubelet_node_status.go:79] "Successfully registered node" node="master-0" Dec 05 10:37:39.908072 master-0 kubenswrapper[10444]: E1205 10:37:39.908024 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"bootstrap-kube-scheduler-master-0\" already exists" pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:37:39.908746 master-0 kubenswrapper[10444]: E1205 10:37:39.908722 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"bootstrap-kube-controller-manager-master-0\" already exists" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:39.966133 master-0 kubenswrapper[10444]: I1205 10:37:39.966074 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-certs\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:37:39.966606 master-0 kubenswrapper[10444]: I1205 10:37:39.966572 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-ssl-certs-host\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:39.966724 master-0 kubenswrapper[10444]: I1205 10:37:39.966709 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-logs\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:39.966869 master-0 kubenswrapper[10444]: I1205 10:37:39.966856 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-secrets\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:37:39.967050 master-0 kubenswrapper[10444]: I1205 10:37:39.967024 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-secrets\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:39.967268 master-0 kubenswrapper[10444]: I1205 10:37:39.967250 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-logs\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:39.967394 master-0 kubenswrapper[10444]: I1205 10:37:39.967381 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-etc-kubernetes-cloud\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:39.967537 master-0 kubenswrapper[10444]: I1205 10:37:39.967517 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-config\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:39.967639 master-0 kubenswrapper[10444]: I1205 10:37:39.967626 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:37:39.967732 master-0 kubenswrapper[10444]: I1205 10:37:39.967719 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-data-dir\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:37:39.967884 master-0 kubenswrapper[10444]: I1205 10:37:39.967870 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-etc-kubernetes-cloud\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:39.968020 master-0 kubenswrapper[10444]: I1205 10:37:39.968007 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-ssl-certs-host\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:39.968187 master-0 kubenswrapper[10444]: I1205 10:37:39.968171 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-secrets\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:39.968298 master-0 kubenswrapper[10444]: I1205 10:37:39.968263 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-config\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:39.968390 master-0 kubenswrapper[10444]: I1205 10:37:39.968374 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-audit-dir\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:39.968495 master-0 kubenswrapper[10444]: I1205 10:37:39.968481 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-logs\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:37:39.968622 master-0 kubenswrapper[10444]: I1205 10:37:39.968609 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-etc-kube\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:37:40.010730 master-0 kubenswrapper[10444]: W1205 10:37:40.010647 10444 warnings.go:70] would violate PodSecurity "restricted:latest": host namespaces (hostNetwork=true), hostPort (container "etcd" uses hostPorts 2379, 2380), privileged (containers "etcdctl", "etcd" must not set securityContext.privileged=true), allowPrivilegeEscalation != false (containers "etcdctl", "etcd" must set securityContext.allowPrivilegeEscalation=false), unrestricted capabilities (containers "etcdctl", "etcd" must set securityContext.capabilities.drop=["ALL"]), restricted volume types (volumes "certs", "data-dir" use restricted volume type "hostPath"), runAsNonRoot != true (pod or containers "etcdctl", "etcd" must set securityContext.runAsNonRoot=true), seccompProfile (pod or containers "etcdctl", "etcd" must set securityContext.seccompProfile.type to "RuntimeDefault" or "Localhost") Dec 05 10:37:40.010730 master-0 kubenswrapper[10444]: E1205 10:37:40.010705 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"bootstrap-kube-apiserver-master-0\" already exists" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.010984 master-0 kubenswrapper[10444]: E1205 10:37:40.010707 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-master-0-master-0\" already exists" pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:37:40.010984 master-0 kubenswrapper[10444]: E1205 10:37:40.010667 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-rbac-proxy-crio-master-0\" already exists" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:37:40.069640 master-0 kubenswrapper[10444]: I1205 10:37:40.069581 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-ssl-certs-host\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:40.069640 master-0 kubenswrapper[10444]: I1205 10:37:40.069633 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-logs\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:40.070070 master-0 kubenswrapper[10444]: I1205 10:37:40.069657 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-secrets\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:37:40.070070 master-0 kubenswrapper[10444]: I1205 10:37:40.069681 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-certs\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:37:40.070070 master-0 kubenswrapper[10444]: I1205 10:37:40.069706 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-logs\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070070 master-0 kubenswrapper[10444]: I1205 10:37:40.069731 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-etc-kubernetes-cloud\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:40.070070 master-0 kubenswrapper[10444]: I1205 10:37:40.069754 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-config\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:40.070070 master-0 kubenswrapper[10444]: I1205 10:37:40.069778 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:37:40.070070 master-0 kubenswrapper[10444]: I1205 10:37:40.069843 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-secrets\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070070 master-0 kubenswrapper[10444]: I1205 10:37:40.070059 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-etc-kubernetes-cloud\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070288 master-0 kubenswrapper[10444]: I1205 10:37:40.070081 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-ssl-certs-host\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070288 master-0 kubenswrapper[10444]: I1205 10:37:40.070104 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-secrets\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:40.070288 master-0 kubenswrapper[10444]: I1205 10:37:40.070125 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-data-dir\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:37:40.070288 master-0 kubenswrapper[10444]: I1205 10:37:40.070143 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-audit-dir\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070288 master-0 kubenswrapper[10444]: I1205 10:37:40.070162 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-logs\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:37:40.070288 master-0 kubenswrapper[10444]: I1205 10:37:40.070189 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-etc-kube\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:37:40.070288 master-0 kubenswrapper[10444]: I1205 10:37:40.070209 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-config\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070288 master-0 kubenswrapper[10444]: I1205 10:37:40.070271 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-config\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070561 master-0 kubenswrapper[10444]: I1205 10:37:40.070334 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-ssl-certs-host\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:40.070561 master-0 kubenswrapper[10444]: I1205 10:37:40.070365 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-logs\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:40.070561 master-0 kubenswrapper[10444]: I1205 10:37:40.070395 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-secrets\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:37:40.070561 master-0 kubenswrapper[10444]: I1205 10:37:40.070429 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-certs\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:37:40.070561 master-0 kubenswrapper[10444]: I1205 10:37:40.070479 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-logs\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070561 master-0 kubenswrapper[10444]: I1205 10:37:40.070521 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-etc-kubernetes-cloud\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:40.070742 master-0 kubenswrapper[10444]: I1205 10:37:40.070588 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-secrets\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070742 master-0 kubenswrapper[10444]: I1205 10:37:40.070620 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-ssl-certs-host\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070742 master-0 kubenswrapper[10444]: I1205 10:37:40.070627 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-etc-kubernetes-cloud\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070742 master-0 kubenswrapper[10444]: I1205 10:37:40.070672 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-data-dir\") pod \"etcd-master-0-master-0\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:37:40.070742 master-0 kubenswrapper[10444]: I1205 10:37:40.070710 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-audit-dir\") pod \"bootstrap-kube-apiserver-master-0\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:40.070742 master-0 kubenswrapper[10444]: I1205 10:37:40.070730 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-logs\") pod \"bootstrap-kube-scheduler-master-0\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:37:40.070908 master-0 kubenswrapper[10444]: I1205 10:37:40.070863 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-etc-kube\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:37:40.070939 master-0 kubenswrapper[10444]: I1205 10:37:40.070908 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:37:40.070970 master-0 kubenswrapper[10444]: I1205 10:37:40.070937 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-config\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:40.071001 master-0 kubenswrapper[10444]: I1205 10:37:40.070975 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-secrets\") pod \"bootstrap-kube-controller-manager-master-0\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:40.525069 master-0 kubenswrapper[10444]: I1205 10:37:40.524981 10444 apiserver.go:52] "Watching apiserver" Dec 05 10:37:40.538226 master-0 kubenswrapper[10444]: I1205 10:37:40.537909 10444 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 10:37:40.539069 master-0 kubenswrapper[10444]: I1205 10:37:40.538996 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t","openshift-multus/multus-admission-controller-7dfc5b745f-67rx7","openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk","openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8","openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c","openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb","openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p","openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m","openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8","openshift-multus/multus-additional-cni-plugins-dms5d","openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp","openshift-ovn-kubernetes/ovnkube-node-rsfjs","openshift-etcd/etcd-master-0-master-0","openshift-ingress-operator/ingress-operator-8649c48786-cgt5x","openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr","openshift-network-node-identity/network-node-identity-ql7j7","openshift-network-operator/network-operator-79767b7ff9-t8j2j","openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr","kube-system/bootstrap-kube-scheduler-master-0","openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27","openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv","openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4","assisted-installer/assisted-installer-controller-pd4q6","openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f","openshift-marketplace/marketplace-operator-f797b99b6-z9qcl","openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd","openshift-multus/multus-lxmgz","openshift-multus/network-metrics-daemon-8gjgm","openshift-controller-manager/controller-manager-77f4fc6d5d-zdn92","openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h","openshift-kube-apiserver/bootstrap-kube-apiserver-master-0","openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr","openshift-machine-config-operator/kube-rbac-proxy-crio-master-0","openshift-network-diagnostics/network-check-target-d6fzk","openshift-network-operator/iptables-alerter-d6wjk","openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn","kube-system/bootstrap-kube-controller-manager-master-0","openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7","openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5","openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v","openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl","openshift-service-ca/service-ca-77c99c46b8-m7zqs"] Dec 05 10:37:40.539264 master-0 kubenswrapper[10444]: I1205 10:37:40.539235 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:37:40.539791 master-0 kubenswrapper[10444]: I1205 10:37:40.539742 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:40.539876 master-0 kubenswrapper[10444]: I1205 10:37:40.539828 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:40.541105 master-0 kubenswrapper[10444]: I1205 10:37:40.540178 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:40.541105 master-0 kubenswrapper[10444]: I1205 10:37:40.540255 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:40.541105 master-0 kubenswrapper[10444]: I1205 10:37:40.541043 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.541272 master-0 kubenswrapper[10444]: I1205 10:37:40.541127 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:40.541341 master-0 kubenswrapper[10444]: I1205 10:37:40.541303 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:40.546323 master-0 kubenswrapper[10444]: I1205 10:37:40.546226 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:40.546638 master-0 kubenswrapper[10444]: I1205 10:37:40.546371 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:40.550456 master-0 kubenswrapper[10444]: I1205 10:37:40.550399 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 10:37:40.550601 master-0 kubenswrapper[10444]: I1205 10:37:40.550482 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 10:37:40.550913 master-0 kubenswrapper[10444]: I1205 10:37:40.550876 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 10:37:40.551325 master-0 kubenswrapper[10444]: I1205 10:37:40.551292 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Dec 05 10:37:40.559166 master-0 kubenswrapper[10444]: I1205 10:37:40.559044 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 10:37:40.559385 master-0 kubenswrapper[10444]: I1205 10:37:40.559188 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Dec 05 10:37:40.559385 master-0 kubenswrapper[10444]: I1205 10:37:40.559065 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.559526 master-0 kubenswrapper[10444]: I1205 10:37:40.559496 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 10:37:40.560000 master-0 kubenswrapper[10444]: I1205 10:37:40.559926 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 10:37:40.560298 master-0 kubenswrapper[10444]: I1205 10:37:40.560142 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 10:37:40.560687 master-0 kubenswrapper[10444]: I1205 10:37:40.560635 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.560994 master-0 kubenswrapper[10444]: I1205 10:37:40.560963 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.561645 master-0 kubenswrapper[10444]: I1205 10:37:40.561586 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-node-tuning-operator"/"node-tuning-operator-tls" Dec 05 10:37:40.561837 master-0 kubenswrapper[10444]: I1205 10:37:40.561609 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-node-tuning-operator"/"performance-addon-operator-webhook-cert" Dec 05 10:37:40.561837 master-0 kubenswrapper[10444]: I1205 10:37:40.561825 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 10:37:40.561937 master-0 kubenswrapper[10444]: I1205 10:37:40.561814 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 10:37:40.561999 master-0 kubenswrapper[10444]: I1205 10:37:40.561980 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 10:37:40.562052 master-0 kubenswrapper[10444]: I1205 10:37:40.562014 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.562052 master-0 kubenswrapper[10444]: I1205 10:37:40.562027 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 10:37:40.562269 master-0 kubenswrapper[10444]: I1205 10:37:40.562237 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 10:37:40.562444 master-0 kubenswrapper[10444]: I1205 10:37:40.562399 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 10:37:40.562729 master-0 kubenswrapper[10444]: I1205 10:37:40.562709 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 10:37:40.562958 master-0 kubenswrapper[10444]: I1205 10:37:40.562940 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 10:37:40.563012 master-0 kubenswrapper[10444]: I1205 10:37:40.562954 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-storage-operator"/"kube-root-ca.crt" Dec 05 10:37:40.563064 master-0 kubenswrapper[10444]: I1205 10:37:40.562710 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.563323 master-0 kubenswrapper[10444]: I1205 10:37:40.563294 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 10:37:40.563804 master-0 kubenswrapper[10444]: I1205 10:37:40.563762 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 10:37:40.565411 master-0 kubenswrapper[10444]: I1205 10:37:40.565365 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:40.566186 master-0 kubenswrapper[10444]: I1205 10:37:40.566152 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:40.566692 master-0 kubenswrapper[10444]: I1205 10:37:40.566654 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-storage-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.566771 master-0 kubenswrapper[10444]: I1205 10:37:40.566730 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 10:37:40.566925 master-0 kubenswrapper[10444]: I1205 10:37:40.566798 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 10:37:40.566925 master-0 kubenswrapper[10444]: I1205 10:37:40.566896 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 10:37:40.567015 master-0 kubenswrapper[10444]: I1205 10:37:40.566985 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 10:37:40.567088 master-0 kubenswrapper[10444]: I1205 10:37:40.566661 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Dec 05 10:37:40.567219 master-0 kubenswrapper[10444]: I1205 10:37:40.567192 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 10:37:40.567516 master-0 kubenswrapper[10444]: I1205 10:37:40.567249 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-olm-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.569554 master-0 kubenswrapper[10444]: I1205 10:37:40.569525 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 10:37:40.569663 master-0 kubenswrapper[10444]: I1205 10:37:40.569635 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.569760 master-0 kubenswrapper[10444]: I1205 10:37:40.569721 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.569925 master-0 kubenswrapper[10444]: I1205 10:37:40.569880 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:37:40.570008 master-0 kubenswrapper[10444]: I1205 10:37:40.569979 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:40.570090 master-0 kubenswrapper[10444]: I1205 10:37:40.570071 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 10:37:40.570283 master-0 kubenswrapper[10444]: I1205 10:37:40.570255 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 10:37:40.570375 master-0 kubenswrapper[10444]: I1205 10:37:40.570356 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 10:37:40.570494 master-0 kubenswrapper[10444]: I1205 10:37:40.570474 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-olm-operator"/"cluster-olm-operator-serving-cert" Dec 05 10:37:40.570550 master-0 kubenswrapper[10444]: I1205 10:37:40.570530 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77f4fc6d5d-zdn92" Dec 05 10:37:40.570594 master-0 kubenswrapper[10444]: I1205 10:37:40.570580 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.570628 master-0 kubenswrapper[10444]: I1205 10:37:40.570594 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 10:37:40.570663 master-0 kubenswrapper[10444]: I1205 10:37:40.570628 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:37:40.570760 master-0 kubenswrapper[10444]: I1205 10:37:40.570739 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 10:37:40.570896 master-0 kubenswrapper[10444]: I1205 10:37:40.570868 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 10:37:40.571334 master-0 kubenswrapper[10444]: I1205 10:37:40.571295 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 10:37:40.571334 master-0 kubenswrapper[10444]: I1205 10:37:40.571324 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 10:37:40.571448 master-0 kubenswrapper[10444]: I1205 10:37:40.571359 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 10:37:40.571693 master-0 kubenswrapper[10444]: I1205 10:37:40.571657 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 10:37:40.571739 master-0 kubenswrapper[10444]: I1205 10:37:40.571666 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.571815 master-0 kubenswrapper[10444]: I1205 10:37:40.571787 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 10:37:40.571867 master-0 kubenswrapper[10444]: I1205 10:37:40.571827 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 10:37:40.571867 master-0 kubenswrapper[10444]: I1205 10:37:40.571333 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 10:37:40.571954 master-0 kubenswrapper[10444]: I1205 10:37:40.571898 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 10:37:40.571954 master-0 kubenswrapper[10444]: I1205 10:37:40.571793 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 10:37:40.572023 master-0 kubenswrapper[10444]: I1205 10:37:40.571917 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 10:37:40.572130 master-0 kubenswrapper[10444]: I1205 10:37:40.572101 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 10:37:40.572247 master-0 kubenswrapper[10444]: I1205 10:37:40.572224 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 10:37:40.572343 master-0 kubenswrapper[10444]: I1205 10:37:40.572321 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 10:37:40.572449 master-0 kubenswrapper[10444]: I1205 10:37:40.572361 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 10:37:40.572510 master-0 kubenswrapper[10444]: I1205 10:37:40.572489 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 10:37:40.572547 master-0 kubenswrapper[10444]: I1205 10:37:40.572522 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 10:37:40.572614 master-0 kubenswrapper[10444]: I1205 10:37:40.572583 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.572647 master-0 kubenswrapper[10444]: I1205 10:37:40.572630 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 10:37:40.572647 master-0 kubenswrapper[10444]: I1205 10:37:40.572633 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 10:37:40.572751 master-0 kubenswrapper[10444]: I1205 10:37:40.572732 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 10:37:40.572822 master-0 kubenswrapper[10444]: I1205 10:37:40.572333 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 10:37:40.572853 master-0 kubenswrapper[10444]: I1205 10:37:40.572822 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 10:37:40.572887 master-0 kubenswrapper[10444]: I1205 10:37:40.572589 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 10:37:40.572987 master-0 kubenswrapper[10444]: I1205 10:37:40.572969 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 10:37:40.573016 master-0 kubenswrapper[10444]: I1205 10:37:40.572982 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"kube-root-ca.crt" Dec 05 10:37:40.573016 master-0 kubenswrapper[10444]: I1205 10:37:40.572995 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 10:37:40.573099 master-0 kubenswrapper[10444]: I1205 10:37:40.569747 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Dec 05 10:37:40.573282 master-0 kubenswrapper[10444]: I1205 10:37:40.573256 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 10:37:40.573424 master-0 kubenswrapper[10444]: I1205 10:37:40.573385 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-ovnkube-config\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:37:40.573501 master-0 kubenswrapper[10444]: I1205 10:37:40.573485 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 10:37:40.573534 master-0 kubenswrapper[10444]: I1205 10:37:40.573508 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-env-overrides\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:37:40.573849 master-0 kubenswrapper[10444]: I1205 10:37:40.573563 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwqkb\" (UniqueName: \"kubernetes.io/projected/1a0235af-2cf2-4ad4-b419-764fb56a0107-kube-api-access-cwqkb\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:37:40.573849 master-0 kubenswrapper[10444]: I1205 10:37:40.573615 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khfxz\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-kube-api-access-khfxz\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:40.573849 master-0 kubenswrapper[10444]: I1205 10:37:40.573681 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-os-release\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.573849 master-0 kubenswrapper[10444]: I1205 10:37:40.573754 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-cni-binary-copy\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.573849 master-0 kubenswrapper[10444]: I1205 10:37:40.573809 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-bound-sa-token\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:40.573987 master-0 kubenswrapper[10444]: I1205 10:37:40.573863 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dtfn\" (UniqueName: \"kubernetes.io/projected/5a1bdc70-6412-47e0-8330-04d796cc8d55-kube-api-access-5dtfn\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:40.573987 master-0 kubenswrapper[10444]: I1205 10:37:40.573921 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:40.574045 master-0 kubenswrapper[10444]: I1205 10:37:40.573980 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-etc-kubernetes\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.574076 master-0 kubenswrapper[10444]: I1205 10:37:40.574041 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-cni-binary-copy\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.574106 master-0 kubenswrapper[10444]: I1205 10:37:40.574036 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/444f8808-e454-4015-9e20-429e715a08c7-config\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:40.574139 master-0 kubenswrapper[10444]: I1205 10:37:40.574120 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lwgq\" (UniqueName: \"kubernetes.io/projected/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-kube-api-access-2lwgq\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:40.574176 master-0 kubenswrapper[10444]: I1205 10:37:40.574158 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:40.574222 master-0 kubenswrapper[10444]: I1205 10:37:40.574189 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11f563d5-89bb-433c-956a-6d5d2492e8f1-kube-api-access\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:40.574222 master-0 kubenswrapper[10444]: I1205 10:37:40.574211 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-k8s-cni-cncf-io\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.574299 master-0 kubenswrapper[10444]: I1205 10:37:40.574234 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operand-assets\" (UniqueName: \"kubernetes.io/empty-dir/49051e6e-5a2f-45c8-bad0-374514a91c07-operand-assets\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:40.574299 master-0 kubenswrapper[10444]: I1205 10:37:40.574257 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11f563d5-89bb-433c-956a-6d5d2492e8f1-serving-cert\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:40.574299 master-0 kubenswrapper[10444]: I1205 10:37:40.574278 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"whereabouts-configmap\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-whereabouts-configmap\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.574299 master-0 kubenswrapper[10444]: I1205 10:37:40.574298 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fncz7\" (UniqueName: \"kubernetes.io/projected/926263c4-ec5b-41cb-9c30-0c88f636035f-kube-api-access-fncz7\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:40.574513 master-0 kubenswrapper[10444]: I1205 10:37:40.574320 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssssf\" (UniqueName: \"kubernetes.io/projected/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-kube-api-access-ssssf\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:40.574513 master-0 kubenswrapper[10444]: I1205 10:37:40.574344 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-service-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:40.574513 master-0 kubenswrapper[10444]: I1205 10:37:40.574373 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkmvj\" (UniqueName: \"kubernetes.io/projected/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-kube-api-access-fkmvj\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:40.574513 master-0 kubenswrapper[10444]: I1205 10:37:40.574395 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c22d947f-a5b6-4f24-b142-dd201c46293b-serving-cert\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:40.574513 master-0 kubenswrapper[10444]: I1205 10:37:40.574416 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:40.574513 master-0 kubenswrapper[10444]: I1205 10:37:40.574463 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-bin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.574513 master-0 kubenswrapper[10444]: I1205 10:37:40.574488 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-service-ca\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.574513 master-0 kubenswrapper[10444]: I1205 10:37:40.574511 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twh85\" (UniqueName: \"kubernetes.io/projected/f7a08359-0379-4364-8b0c-ddb58ff605f4-kube-api-access-twh85\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574534 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-config\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574558 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574580 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-serving-cert\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574604 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-service-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574626 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-binary-copy\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574648 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574673 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfz87\" (UniqueName: \"kubernetes.io/projected/b760849c-8d83-47da-8677-68445c143bef-kube-api-access-jfz87\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574697 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c22d947f-a5b6-4f24-b142-dd201c46293b-available-featuregates\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574721 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98qt8\" (UniqueName: \"kubernetes.io/projected/c22d947f-a5b6-4f24-b142-dd201c46293b-kube-api-access-98qt8\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574746 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-trusted-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574767 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-bound-sa-token\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574787 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fbss\" (UniqueName: \"kubernetes.io/projected/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-kube-api-access-2fbss\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:40.574829 master-0 kubenswrapper[10444]: I1205 10:37:40.574794 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/444f8808-e454-4015-9e20-429e715a08c7-config\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:40.575400 master-0 kubenswrapper[10444]: I1205 10:37:40.575014 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operand-assets\" (UniqueName: \"kubernetes.io/empty-dir/49051e6e-5a2f-45c8-bad0-374514a91c07-operand-assets\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:40.575400 master-0 kubenswrapper[10444]: I1205 10:37:40.575262 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f76d12f-5406-47e2-8337-2f50e35376d6-serving-cert\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:40.575400 master-0 kubenswrapper[10444]: I1205 10:37:40.575377 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11f563d5-89bb-433c-956a-6d5d2492e8f1-serving-cert\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:40.575766 master-0 kubenswrapper[10444]: I1205 10:37:40.574808 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f76d12f-5406-47e2-8337-2f50e35376d6-serving-cert\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:40.575846 master-0 kubenswrapper[10444]: I1205 10:37:40.575802 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.575910 master-0 kubenswrapper[10444]: I1205 10:37:40.575877 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:40.575976 master-0 kubenswrapper[10444]: I1205 10:37:40.575938 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-netns\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.576008 master-0 kubenswrapper[10444]: I1205 10:37:40.575953 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c22d947f-a5b6-4f24-b142-dd201c46293b-serving-cert\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:40.576056 master-0 kubenswrapper[10444]: I1205 10:37:40.576021 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4825316a-ea9f-4d3d-838b-fa809a6e49c7-serving-cert\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:40.576158 master-0 kubenswrapper[10444]: I1205 10:37:40.576110 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkqz7\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-kube-api-access-vkqz7\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:40.576229 master-0 kubenswrapper[10444]: I1205 10:37:40.576194 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjrgm\" (UniqueName: \"kubernetes.io/projected/87909f47-f2d7-46f8-a1c8-27336cdcce5d-kube-api-access-rjrgm\") pod \"csi-snapshot-controller-operator-6bc8656fdc-vd94f\" (UID: \"87909f47-f2d7-46f8-a1c8-27336cdcce5d\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:40.576260 master-0 kubenswrapper[10444]: I1205 10:37:40.576235 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-service-ca\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.576302 master-0 kubenswrapper[10444]: I1205 10:37:40.576265 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-multus-daemon-config\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.576354 master-0 kubenswrapper[10444]: I1205 10:37:40.576321 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:40.576426 master-0 kubenswrapper[10444]: I1205 10:37:40.576388 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrdsv\" (UniqueName: \"kubernetes.io/projected/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-kube-api-access-jrdsv\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.576546 master-0 kubenswrapper[10444]: I1205 10:37:40.576511 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-socket-dir-parent\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.576601 master-0 kubenswrapper[10444]: I1205 10:37:40.576547 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-config\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:40.576601 master-0 kubenswrapper[10444]: I1205 10:37:40.576584 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/8c649a16-c187-412e-b5da-62a00bee38ab-kube-api-access-d8pv2\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:40.576685 master-0 kubenswrapper[10444]: I1205 10:37:40.576640 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f563d5-89bb-433c-956a-6d5d2492e8f1-config\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:40.576732 master-0 kubenswrapper[10444]: I1205 10:37:40.576704 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:40.576732 master-0 kubenswrapper[10444]: I1205 10:37:40.576710 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-hostroot\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.576810 master-0 kubenswrapper[10444]: I1205 10:37:40.576771 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4825316a-ea9f-4d3d-838b-fa809a6e49c7-config\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:40.576854 master-0 kubenswrapper[10444]: I1205 10:37:40.576828 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-serving-cert\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:40.576932 master-0 kubenswrapper[10444]: I1205 10:37:40.576886 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-telemetry-config\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:40.576993 master-0 kubenswrapper[10444]: I1205 10:37:40.576893 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-serving-cert\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:40.576993 master-0 kubenswrapper[10444]: I1205 10:37:40.576974 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-kubelet\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.577073 master-0 kubenswrapper[10444]: I1205 10:37:40.577034 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:40.577073 master-0 kubenswrapper[10444]: I1205 10:37:40.577062 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-service-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:40.577157 master-0 kubenswrapper[10444]: I1205 10:37:40.577088 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn7rj\" (UniqueName: \"kubernetes.io/projected/eb290494-a456-4f0e-9afc-f20abab1a1bf-kube-api-access-qn7rj\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:40.577157 master-0 kubenswrapper[10444]: I1205 10:37:40.577141 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-client\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:40.577236 master-0 kubenswrapper[10444]: I1205 10:37:40.577178 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/444f8808-e454-4015-9e20-429e715a08c7-kube-api-access\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:40.577236 master-0 kubenswrapper[10444]: I1205 10:37:40.577216 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a0235af-2cf2-4ad4-b419-764fb56a0107-metrics-tls\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:37:40.577321 master-0 kubenswrapper[10444]: I1205 10:37:40.577254 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-serving-cert\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:40.577321 master-0 kubenswrapper[10444]: I1205 10:37:40.577277 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-binary-copy\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.577404 master-0 kubenswrapper[10444]: I1205 10:37:40.575720 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-service-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:40.577674 master-0 kubenswrapper[10444]: I1205 10:37:40.577640 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4825316a-ea9f-4d3d-838b-fa809a6e49c7-config\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:40.577730 master-0 kubenswrapper[10444]: I1205 10:37:40.577626 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c22d947f-a5b6-4f24-b142-dd201c46293b-available-featuregates\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:40.577774 master-0 kubenswrapper[10444]: I1205 10:37:40.577360 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-serving-cert\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:40.578162 master-0 kubenswrapper[10444]: I1205 10:37:40.578121 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-olm-operator"/"kube-root-ca.crt" Dec 05 10:37:40.578293 master-0 kubenswrapper[10444]: I1205 10:37:40.578228 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-serving-cert\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:40.578356 master-0 kubenswrapper[10444]: I1205 10:37:40.578332 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-client\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:40.578423 master-0 kubenswrapper[10444]: I1205 10:37:40.578385 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-telemetry-config\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:40.578495 master-0 kubenswrapper[10444]: I1205 10:37:40.578447 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-serving-cert\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:40.578495 master-0 kubenswrapper[10444]: I1205 10:37:40.578465 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-trusted-ca\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:40.578568 master-0 kubenswrapper[10444]: I1205 10:37:40.578501 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-cnibin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.578568 master-0 kubenswrapper[10444]: I1205 10:37:40.578525 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-multus\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.578729 master-0 kubenswrapper[10444]: I1205 10:37:40.578687 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-serving-cert\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:40.578778 master-0 kubenswrapper[10444]: I1205 10:37:40.578741 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f563d5-89bb-433c-956a-6d5d2492e8f1-config\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:40.578816 master-0 kubenswrapper[10444]: I1205 10:37:40.578132 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4825316a-ea9f-4d3d-838b-fa809a6e49c7-serving-cert\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:40.579070 master-0 kubenswrapper[10444]: I1205 10:37:40.579037 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-multus-daemon-config\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.580849 master-0 kubenswrapper[10444]: I1205 10:37:40.580788 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a0235af-2cf2-4ad4-b419-764fb56a0107-metrics-tls\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:37:40.581148 master-0 kubenswrapper[10444]: I1205 10:37:40.581113 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 10:37:40.586346 master-0 kubenswrapper[10444]: I1205 10:37:40.586271 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:40.586346 master-0 kubenswrapper[10444]: I1205 10:37:40.586343 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vp9b\" (UniqueName: \"kubernetes.io/projected/49051e6e-5a2f-45c8-bad0-374514a91c07-kube-api-access-6vp9b\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:40.586643 master-0 kubenswrapper[10444]: I1205 10:37:40.586370 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-config\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:40.586935 master-0 kubenswrapper[10444]: I1205 10:37:40.586889 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"whereabouts-config" Dec 05 10:37:40.587030 master-0 kubenswrapper[10444]: I1205 10:37:40.586571 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7b29f89-e42d-4e53-ad14-05efdce933f0-trusted-ca\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:40.587101 master-0 kubenswrapper[10444]: I1205 10:37:40.587066 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-system-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.587101 master-0 kubenswrapper[10444]: I1205 10:37:40.587088 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.587222 master-0 kubenswrapper[10444]: I1205 10:37:40.587196 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f76d12f-5406-47e2-8337-2f50e35376d6-config\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:40.587287 master-0 kubenswrapper[10444]: I1205 10:37:40.587227 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:40.587287 master-0 kubenswrapper[10444]: I1205 10:37:40.587284 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvnxf\" (UniqueName: \"kubernetes.io/projected/6a9f011f-36f1-4308-a365-69425c186c7f-kube-api-access-mvnxf\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.587379 master-0 kubenswrapper[10444]: I1205 10:37:40.587302 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-kube-api-access\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.587758 master-0 kubenswrapper[10444]: I1205 10:37:40.587731 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e69ce9e-4e6f-4015-9ba6-5a7942570190-config\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:40.587853 master-0 kubenswrapper[10444]: I1205 10:37:40.587635 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f76d12f-5406-47e2-8337-2f50e35376d6-config\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:40.588269 master-0 kubenswrapper[10444]: I1205 10:37:40.588227 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e69ce9e-4e6f-4015-9ba6-5a7942570190-config\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:40.588327 master-0 kubenswrapper[10444]: I1205 10:37:40.586637 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-config\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:40.588419 master-0 kubenswrapper[10444]: I1205 10:37:40.588384 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eb290494-a456-4f0e-9afc-f20abab1a1bf-trusted-ca\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:40.588474 master-0 kubenswrapper[10444]: I1205 10:37:40.588422 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e69ce9e-4e6f-4015-9ba6-5a7942570190-serving-cert\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:40.588474 master-0 kubenswrapper[10444]: I1205 10:37:40.588453 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:40.588531 master-0 kubenswrapper[10444]: I1205 10:37:40.588481 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-ssl-certs\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.588531 master-0 kubenswrapper[10444]: I1205 10:37:40.588502 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-config\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:40.588531 master-0 kubenswrapper[10444]: I1205 10:37:40.588522 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzwfq\" (UniqueName: \"kubernetes.io/projected/6f76d12f-5406-47e2-8337-2f50e35376d6-kube-api-access-fzwfq\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:40.588643 master-0 kubenswrapper[10444]: I1205 10:37:40.588545 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-olm-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/49051e6e-5a2f-45c8-bad0-374514a91c07-cluster-olm-operator-serving-cert\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:40.588643 master-0 kubenswrapper[10444]: I1205 10:37:40.588567 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4825316a-ea9f-4d3d-838b-fa809a6e49c7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:40.588643 master-0 kubenswrapper[10444]: I1205 10:37:40.588591 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2vd4\" (UniqueName: \"kubernetes.io/projected/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-kube-api-access-h2vd4\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:40.588735 master-0 kubenswrapper[10444]: I1205 10:37:40.588619 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-config\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:40.588735 master-0 kubenswrapper[10444]: I1205 10:37:40.588683 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kb5c\" (UniqueName: \"kubernetes.io/projected/8d76404b-6d62-4a61-b6f6-0c8073eba198-kube-api-access-7kb5c\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:40.588790 master-0 kubenswrapper[10444]: I1205 10:37:40.588767 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b760849c-8d83-47da-8677-68445c143bef-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:37:40.588836 master-0 kubenswrapper[10444]: I1205 10:37:40.588819 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-system-cni-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.588873 master-0 kubenswrapper[10444]: I1205 10:37:40.588849 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-multus-certs\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.588906 master-0 kubenswrapper[10444]: I1205 10:37:40.588865 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 10:37:40.588906 master-0 kubenswrapper[10444]: I1205 10:37:40.588874 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-profile-collector-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:40.588960 master-0 kubenswrapper[10444]: I1205 10:37:40.588906 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.588960 master-0 kubenswrapper[10444]: I1205 10:37:40.588928 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.588960 master-0 kubenswrapper[10444]: I1205 10:37:40.588950 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-conf-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.589051 master-0 kubenswrapper[10444]: I1205 10:37:40.588972 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22676fac-b770-4937-9bee-7478bd1babb7-trusted-ca\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:40.589051 master-0 kubenswrapper[10444]: I1205 10:37:40.588993 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:40.589051 master-0 kubenswrapper[10444]: I1205 10:37:40.589014 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cnibin\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.589051 master-0 kubenswrapper[10444]: I1205 10:37:40.589039 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv4zs\" (UniqueName: \"kubernetes.io/projected/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-kube-api-access-jv4zs\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:40.589199 master-0 kubenswrapper[10444]: I1205 10:37:40.589063 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.589199 master-0 kubenswrapper[10444]: I1205 10:37:40.589083 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-os-release\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.589199 master-0 kubenswrapper[10444]: I1205 10:37:40.589107 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wfsv\" (UniqueName: \"kubernetes.io/projected/1e69ce9e-4e6f-4015-9ba6-5a7942570190-kube-api-access-7wfsv\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:40.589199 master-0 kubenswrapper[10444]: I1205 10:37:40.589128 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/1a0235af-2cf2-4ad4-b419-764fb56a0107-host-etc-kube\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:37:40.589199 master-0 kubenswrapper[10444]: I1205 10:37:40.589151 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/444f8808-e454-4015-9e20-429e715a08c7-serving-cert\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:40.589512 master-0 kubenswrapper[10444]: I1205 10:37:40.589266 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-config\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:40.589512 master-0 kubenswrapper[10444]: I1205 10:37:40.589338 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/444f8808-e454-4015-9e20-429e715a08c7-serving-cert\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:40.589512 master-0 kubenswrapper[10444]: I1205 10:37:40.589478 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-olm-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/49051e6e-5a2f-45c8-bad0-374514a91c07-cluster-olm-operator-serving-cert\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:40.589635 master-0 kubenswrapper[10444]: I1205 10:37:40.589522 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.589750 master-0 kubenswrapper[10444]: I1205 10:37:40.589705 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e69ce9e-4e6f-4015-9ba6-5a7942570190-serving-cert\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:40.589750 master-0 kubenswrapper[10444]: I1205 10:37:40.589712 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-config\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:40.589842 master-0 kubenswrapper[10444]: I1205 10:37:40.589811 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-profile-collector-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:40.608317 master-0 kubenswrapper[10444]: I1205 10:37:40.608241 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 10:37:40.608523 master-0 kubenswrapper[10444]: I1205 10:37:40.608330 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"whereabouts-configmap\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-whereabouts-configmap\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.610154 master-0 kubenswrapper[10444]: I1205 10:37:40.610116 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 10:37:40.610992 master-0 kubenswrapper[10444]: I1205 10:37:40.610949 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 10:37:40.611471 master-0 kubenswrapper[10444]: I1205 10:37:40.611397 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 10:37:40.611677 master-0 kubenswrapper[10444]: I1205 10:37:40.611635 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 10:37:40.611906 master-0 kubenswrapper[10444]: I1205 10:37:40.611822 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-trusted-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:40.611980 master-0 kubenswrapper[10444]: I1205 10:37:40.611919 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 10:37:40.612153 master-0 kubenswrapper[10444]: I1205 10:37:40.612121 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 10:37:40.612411 master-0 kubenswrapper[10444]: I1205 10:37:40.612372 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7b29f89-e42d-4e53-ad14-05efdce933f0-trusted-ca\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:40.612538 master-0 kubenswrapper[10444]: I1205 10:37:40.612479 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 10:37:40.612593 master-0 kubenswrapper[10444]: I1205 10:37:40.612538 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 10:37:40.612723 master-0 kubenswrapper[10444]: I1205 10:37:40.612703 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:37:40.612782 master-0 kubenswrapper[10444]: I1205 10:37:40.612691 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 10:37:40.612846 master-0 kubenswrapper[10444]: I1205 10:37:40.612824 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 10:37:40.612915 master-0 kubenswrapper[10444]: I1205 10:37:40.612881 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 10:37:40.613145 master-0 kubenswrapper[10444]: I1205 10:37:40.613077 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 10:37:40.613413 master-0 kubenswrapper[10444]: I1205 10:37:40.613366 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 10:37:40.614724 master-0 kubenswrapper[10444]: I1205 10:37:40.614681 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-ovnkube-config\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:37:40.614805 master-0 kubenswrapper[10444]: I1205 10:37:40.614782 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 10:37:40.614965 master-0 kubenswrapper[10444]: I1205 10:37:40.614928 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 10:37:40.616863 master-0 kubenswrapper[10444]: I1205 10:37:40.616821 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 10:37:40.619326 master-0 kubenswrapper[10444]: I1205 10:37:40.619267 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b760849c-8d83-47da-8677-68445c143bef-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:37:40.619862 master-0 kubenswrapper[10444]: I1205 10:37:40.619805 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22676fac-b770-4937-9bee-7478bd1babb7-trusted-ca\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:40.620681 master-0 kubenswrapper[10444]: I1205 10:37:40.620632 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 10:37:40.621193 master-0 kubenswrapper[10444]: I1205 10:37:40.621149 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"trusted-ca" Dec 05 10:37:40.621267 master-0 kubenswrapper[10444]: I1205 10:37:40.621206 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77f4fc6d5d-zdn92" Dec 05 10:37:40.627722 master-0 kubenswrapper[10444]: I1205 10:37:40.627682 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 10:37:40.629953 master-0 kubenswrapper[10444]: I1205 10:37:40.629905 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eb290494-a456-4f0e-9afc-f20abab1a1bf-trusted-ca\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:40.630077 master-0 kubenswrapper[10444]: I1205 10:37:40.630037 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-trusted-ca\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:40.647785 master-0 kubenswrapper[10444]: I1205 10:37:40.647636 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 10:37:40.648150 master-0 kubenswrapper[10444]: I1205 10:37:40.648115 10444 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Dec 05 10:37:40.656061 master-0 kubenswrapper[10444]: I1205 10:37:40.655983 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-env-overrides\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:37:40.667551 master-0 kubenswrapper[10444]: I1205 10:37:40.667295 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 10:37:40.679705 master-0 kubenswrapper[10444]: I1205 10:37:40.679655 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77f4fc6d5d-zdn92" Dec 05 10:37:40.688257 master-0 kubenswrapper[10444]: I1205 10:37:40.688063 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 10:37:40.689696 master-0 kubenswrapper[10444]: I1205 10:37:40.689511 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/283122ba-be1c-4516-bd0f-df41c13c098b-host-slash\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:40.689696 master-0 kubenswrapper[10444]: I1205 10:37:40.689564 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:40.689696 master-0 kubenswrapper[10444]: I1205 10:37:40.689597 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-cnibin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.689696 master-0 kubenswrapper[10444]: I1205 10:37:40.689619 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-multus\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.689696 master-0 kubenswrapper[10444]: I1205 10:37:40.689640 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-system-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.689932 master-0 kubenswrapper[10444]: E1205 10:37:40.689771 10444 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: secret "multus-admission-controller-secret" not found Dec 05 10:37:40.689932 master-0 kubenswrapper[10444]: I1205 10:37:40.689809 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-multus\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.689932 master-0 kubenswrapper[10444]: E1205 10:37:40.689865 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs podName:8d76404b-6d62-4a61-b6f6-0c8073eba198 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.189838772 +0000 UTC m=+1.781751429 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs") pod "multus-admission-controller-7dfc5b745f-67rx7" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198") : secret "multus-admission-controller-secret" not found Dec 05 10:37:40.689932 master-0 kubenswrapper[10444]: I1205 10:37:40.689888 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-system-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.690073 master-0 kubenswrapper[10444]: I1205 10:37:40.689946 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-cnibin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.690073 master-0 kubenswrapper[10444]: I1205 10:37:40.689995 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.690073 master-0 kubenswrapper[10444]: I1205 10:37:40.690032 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:40.690073 master-0 kubenswrapper[10444]: I1205 10:37:40.690057 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-slash\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.690319 master-0 kubenswrapper[10444]: I1205 10:37:40.690203 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.690319 master-0 kubenswrapper[10444]: E1205 10:37:40.690237 10444 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:37:40.690411 master-0 kubenswrapper[10444]: I1205 10:37:40.690262 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:40.690411 master-0 kubenswrapper[10444]: E1205 10:37:40.690335 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.190309615 +0000 UTC m=+1.782222262 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:37:40.690411 master-0 kubenswrapper[10444]: I1205 10:37:40.690374 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-kubelet\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.690411 master-0 kubenswrapper[10444]: E1205 10:37:40.690402 10444 secret.go:189] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: secret "marketplace-operator-metrics" not found Dec 05 10:37:40.690626 master-0 kubenswrapper[10444]: I1205 10:37:40.690414 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-ssl-certs\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.690626 master-0 kubenswrapper[10444]: I1205 10:37:40.690467 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:40.690626 master-0 kubenswrapper[10444]: E1205 10:37:40.690481 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics podName:926263c4-ec5b-41cb-9c30-0c88f636035f nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.190468249 +0000 UTC m=+1.782380816 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics") pod "marketplace-operator-f797b99b6-z9qcl" (UID: "926263c4-ec5b-41cb-9c30-0c88f636035f") : secret "marketplace-operator-metrics" not found Dec 05 10:37:40.690626 master-0 kubenswrapper[10444]: I1205 10:37:40.690508 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml74w\" (UniqueName: \"kubernetes.io/projected/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f-kube-api-access-ml74w\") pod \"csi-snapshot-controller-6b958b6f94-lgn6v\" (UID: \"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" Dec 05 10:37:40.690626 master-0 kubenswrapper[10444]: I1205 10:37:40.690569 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdsh9\" (UniqueName: \"kubernetes.io/projected/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a-kube-api-access-pdsh9\") pod \"migrator-74b7b57c65-sfvzd\" (UID: \"3c4f767d-65f9-434b-8ddd-ceb0b91ab99a\") " pod="openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd" Dec 05 10:37:40.690626 master-0 kubenswrapper[10444]: I1205 10:37:40.690607 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-netns\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.690626 master-0 kubenswrapper[10444]: I1205 10:37:40.690627 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-systemd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.690882 master-0 kubenswrapper[10444]: I1205 10:37:40.690649 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-log-socket\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.690882 master-0 kubenswrapper[10444]: E1205 10:37:40.690661 10444 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Dec 05 10:37:40.690882 master-0 kubenswrapper[10444]: E1205 10:37:40.690695 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.190685205 +0000 UTC m=+1.782597782 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : secret "metrics-daemon-secret" not found Dec 05 10:37:40.690882 master-0 kubenswrapper[10444]: I1205 10:37:40.690718 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-system-cni-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.690882 master-0 kubenswrapper[10444]: I1205 10:37:40.690738 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.690882 master-0 kubenswrapper[10444]: I1205 10:37:40.690754 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-conf-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.690882 master-0 kubenswrapper[10444]: I1205 10:37:40.690772 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-multus-certs\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.690882 master-0 kubenswrapper[10444]: I1205 10:37:40.690799 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.690882 master-0 kubenswrapper[10444]: I1205 10:37:40.690819 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:40.690882 master-0 kubenswrapper[10444]: I1205 10:37:40.690838 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/283122ba-be1c-4516-bd0f-df41c13c098b-iptables-alerter-script\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:40.690882 master-0 kubenswrapper[10444]: I1205 10:37:40.690854 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cnibin\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: I1205 10:37:40.690945 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/1a0235af-2cf2-4ad4-b419-764fb56a0107-host-etc-kube\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: I1205 10:37:40.690977 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-cvo-updatepayloads\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: I1205 10:37:40.690993 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cnibin\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: I1205 10:37:40.690999 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/1a0235af-2cf2-4ad4-b419-764fb56a0107-host-etc-kube\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: I1205 10:37:40.691008 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-ssl-certs\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: I1205 10:37:40.691028 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-conf-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: I1205 10:37:40.691041 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-system-cni-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: I1205 10:37:40.691083 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: I1205 10:37:40.691089 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-os-release\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: I1205 10:37:40.691112 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-multus-certs\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: E1205 10:37:40.691129 10444 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/performance-addon-operator-webhook-cert: secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: I1205 10:37:40.691219 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-os-release\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.691373 master-0 kubenswrapper[10444]: E1205 10:37:40.691257 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.191246281 +0000 UTC m=+1.783158958 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:40.691784 master-0 kubenswrapper[10444]: I1205 10:37:40.691460 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-os-release\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.691784 master-0 kubenswrapper[10444]: I1205 10:37:40.691510 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/495ba1ea-f844-43ec-8be7-47e738f5428a-ovn-node-metrics-cert\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.691784 master-0 kubenswrapper[10444]: I1205 10:37:40.691552 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-script-lib\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.691784 master-0 kubenswrapper[10444]: I1205 10:37:40.691602 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:40.691784 master-0 kubenswrapper[10444]: I1205 10:37:40.691635 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-etc-kubernetes\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.691784 master-0 kubenswrapper[10444]: I1205 10:37:40.691695 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-ovn\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.691784 master-0 kubenswrapper[10444]: I1205 10:37:40.691728 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:40.692012 master-0 kubenswrapper[10444]: I1205 10:37:40.691801 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:40.692012 master-0 kubenswrapper[10444]: I1205 10:37:40.691823 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-script-lib\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.692012 master-0 kubenswrapper[10444]: I1205 10:37:40.691829 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-bin\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.692012 master-0 kubenswrapper[10444]: E1205 10:37:40.691870 10444 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: secret "olm-operator-serving-cert" not found Dec 05 10:37:40.692012 master-0 kubenswrapper[10444]: I1205 10:37:40.691872 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.692012 master-0 kubenswrapper[10444]: E1205 10:37:40.691923 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert podName:46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.19191201 +0000 UTC m=+1.783824677 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert") pod "olm-operator-7cd7dbb44c-d25sk" (UID: "46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082") : secret "olm-operator-serving-cert" not found Dec 05 10:37:40.692012 master-0 kubenswrapper[10444]: I1205 10:37:40.691948 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-etc-kubernetes\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.692012 master-0 kubenswrapper[10444]: E1205 10:37:40.691986 10444 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:40.692012 master-0 kubenswrapper[10444]: I1205 10:37:40.692005 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/495ba1ea-f844-43ec-8be7-47e738f5428a-ovn-node-metrics-cert\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.692305 master-0 kubenswrapper[10444]: E1205 10:37:40.692016 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls podName:a722cda9-29a0-4b7f-8e1d-9a8950ed765a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.192005352 +0000 UTC m=+1.783918039 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-7ff994598c-kq8qr" (UID: "a722cda9-29a0-4b7f-8e1d-9a8950ed765a") : secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:40.692305 master-0 kubenswrapper[10444]: I1205 10:37:40.692047 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-env-overrides\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.692305 master-0 kubenswrapper[10444]: I1205 10:37:40.692062 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-os-release\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.692305 master-0 kubenswrapper[10444]: I1205 10:37:40.692095 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjgc4\" (UniqueName: \"kubernetes.io/projected/495ba1ea-f844-43ec-8be7-47e738f5428a-kube-api-access-tjgc4\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.692305 master-0 kubenswrapper[10444]: I1205 10:37:40.692122 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-k8s-cni-cncf-io\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.692305 master-0 kubenswrapper[10444]: I1205 10:37:40.692186 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwpcs\" (UniqueName: \"kubernetes.io/projected/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-kube-api-access-fwpcs\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:40.692305 master-0 kubenswrapper[10444]: I1205 10:37:40.692189 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-env-overrides\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.692305 master-0 kubenswrapper[10444]: I1205 10:37:40.692262 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-key\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:37:40.692305 master-0 kubenswrapper[10444]: I1205 10:37:40.692285 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-env-overrides\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:37:40.692644 master-0 kubenswrapper[10444]: I1205 10:37:40.692330 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-bin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.692644 master-0 kubenswrapper[10444]: I1205 10:37:40.692384 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-k8s-cni-cncf-io\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.692644 master-0 kubenswrapper[10444]: I1205 10:37:40.692410 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-bin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.692644 master-0 kubenswrapper[10444]: I1205 10:37:40.692474 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:40.692644 master-0 kubenswrapper[10444]: E1205 10:37:40.692521 10444 secret.go:189] Couldn't get secret openshift-image-registry/image-registry-operator-tls: secret "image-registry-operator-tls" not found Dec 05 10:37:40.692644 master-0 kubenswrapper[10444]: E1205 10:37:40.692548 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls podName:f7b29f89-e42d-4e53-ad14-05efdce933f0 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.192539248 +0000 UTC m=+1.784451815 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls") pod "cluster-image-registry-operator-6fb9f88b7-f29mb" (UID: "f7b29f89-e42d-4e53-ad14-05efdce933f0") : secret "image-registry-operator-tls" not found Dec 05 10:37:40.692644 master-0 kubenswrapper[10444]: I1205 10:37:40.692579 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-node-log\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.692644 master-0 kubenswrapper[10444]: I1205 10:37:40.692633 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-etc-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.692953 master-0 kubenswrapper[10444]: I1205 10:37:40.692674 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-webhook-cert\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:37:40.692953 master-0 kubenswrapper[10444]: I1205 10:37:40.692693 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-env-overrides\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:37:40.692953 master-0 kubenswrapper[10444]: I1205 10:37:40.692869 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.692953 master-0 kubenswrapper[10444]: I1205 10:37:40.692923 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-cabundle\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.692967 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: E1205 10:37:40.693036 10444 secret.go:189] Couldn't get secret openshift-ingress-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693049 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-webhook-cert\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: E1205 10:37:40.693074 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls podName:22676fac-b770-4937-9bee-7478bd1babb7 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.193060802 +0000 UTC m=+1.784973369 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls") pod "ingress-operator-8649c48786-cgt5x" (UID: "22676fac-b770-4937-9bee-7478bd1babb7") : secret "metrics-tls" not found Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693090 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-config\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693106 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-netd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693124 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-netns\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693154 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693172 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693192 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693218 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-netns\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693220 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-ovnkube-identity-cm\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693257 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693260 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693292 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-socket-dir-parent\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693319 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-config\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: E1205 10:37:40.693298 10444 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: secret "package-server-manager-serving-cert" not found Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: E1205 10:37:40.693353 10444 secret.go:189] Couldn't get secret openshift-dns-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693359 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn6sc\" (UniqueName: \"kubernetes.io/projected/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-kube-api-access-nn6sc\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693400 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhjhk\" (UniqueName: \"kubernetes.io/projected/283122ba-be1c-4516-bd0f-df41c13c098b-kube-api-access-vhjhk\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693456 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-hostroot\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693481 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693512 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-kubelet\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693520 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-config\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693538 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-var-lib-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693553 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-socket-dir-parent\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693564 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg9zq\" (UniqueName: \"kubernetes.io/projected/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-kube-api-access-wg9zq\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: E1205 10:37:40.693583 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert podName:8c649a16-c187-412e-b5da-62a00bee38ab nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.193574917 +0000 UTC m=+1.785487484 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert") pod "package-server-manager-67477646d4-nm8cn" (UID: "8c649a16-c187-412e-b5da-62a00bee38ab") : secret "package-server-manager-serving-cert" not found Dec 05 10:37:40.693583 master-0 kubenswrapper[10444]: I1205 10:37:40.693607 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-systemd-units\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.694715 master-0 kubenswrapper[10444]: E1205 10:37:40.693693 10444 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/node-tuning-operator-tls: secret "node-tuning-operator-tls" not found Dec 05 10:37:40.694715 master-0 kubenswrapper[10444]: E1205 10:37:40.693726 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.193715171 +0000 UTC m=+1.785627848 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-tuning-operator-tls" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "node-tuning-operator-tls" not found Dec 05 10:37:40.694715 master-0 kubenswrapper[10444]: I1205 10:37:40.693730 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-kubelet\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.694715 master-0 kubenswrapper[10444]: I1205 10:37:40.693773 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-hostroot\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:40.694715 master-0 kubenswrapper[10444]: E1205 10:37:40.693815 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls podName:5a1bdc70-6412-47e0-8330-04d796cc8d55 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.193807564 +0000 UTC m=+1.785720261 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls") pod "dns-operator-7c56cf9b74-x6t9h" (UID: "5a1bdc70-6412-47e0-8330-04d796cc8d55") : secret "metrics-tls" not found Dec 05 10:37:40.694715 master-0 kubenswrapper[10444]: I1205 10:37:40.693870 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-ovnkube-identity-cm\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:37:40.708580 master-0 kubenswrapper[10444]: I1205 10:37:40.708504 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 10:37:40.726900 master-0 kubenswrapper[10444]: I1205 10:37:40.726821 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:40.728658 master-0 kubenswrapper[10444]: I1205 10:37:40.728576 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 10:37:40.747275 master-0 kubenswrapper[10444]: I1205 10:37:40.747191 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 10:37:40.767655 master-0 kubenswrapper[10444]: I1205 10:37:40.767611 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 10:37:40.787960 master-0 kubenswrapper[10444]: I1205 10:37:40.787890 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 10:37:40.794134 master-0 kubenswrapper[10444]: I1205 10:37:40.794060 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-kubelet\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.794277 master-0 kubenswrapper[10444]: I1205 10:37:40.794207 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-kubelet\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.794366 master-0 kubenswrapper[10444]: I1205 10:37:40.794324 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:40.794482 master-0 kubenswrapper[10444]: E1205 10:37:40.794450 10444 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: configmap "client-ca" not found Dec 05 10:37:40.794530 master-0 kubenswrapper[10444]: E1205 10:37:40.794521 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca podName:98fd575c-ffc4-44c2-9947-c18d3f53e2c9 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.294505181 +0000 UTC m=+1.886417738 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca") pod "route-controller-manager-76d4564964-xm2tr" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9") : configmap "client-ca" not found Dec 05 10:37:40.794699 master-0 kubenswrapper[10444]: I1205 10:37:40.794659 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-netns\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.794746 master-0 kubenswrapper[10444]: I1205 10:37:40.794710 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-systemd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.794746 master-0 kubenswrapper[10444]: I1205 10:37:40.794731 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-log-socket\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.794746 master-0 kubenswrapper[10444]: I1205 10:37:40.794736 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-netns\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.794862 master-0 kubenswrapper[10444]: I1205 10:37:40.794786 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-log-socket\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.794862 master-0 kubenswrapper[10444]: I1205 10:37:40.794816 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-systemd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.794862 master-0 kubenswrapper[10444]: I1205 10:37:40.794856 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-ovn\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.794972 master-0 kubenswrapper[10444]: I1205 10:37:40.794887 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-bin\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.794972 master-0 kubenswrapper[10444]: I1205 10:37:40.794910 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795046 master-0 kubenswrapper[10444]: I1205 10:37:40.794966 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwpcs\" (UniqueName: \"kubernetes.io/projected/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-kube-api-access-fwpcs\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:40.795046 master-0 kubenswrapper[10444]: I1205 10:37:40.794976 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-bin\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795046 master-0 kubenswrapper[10444]: I1205 10:37:40.794992 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-key\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:37:40.795167 master-0 kubenswrapper[10444]: I1205 10:37:40.795061 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795209 master-0 kubenswrapper[10444]: I1205 10:37:40.795184 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-ovn\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795209 master-0 kubenswrapper[10444]: I1205 10:37:40.795188 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-node-log\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795295 master-0 kubenswrapper[10444]: I1205 10:37:40.795242 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-etc-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795295 master-0 kubenswrapper[10444]: I1205 10:37:40.795277 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-node-log\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795295 master-0 kubenswrapper[10444]: I1205 10:37:40.795287 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-etc-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795415 master-0 kubenswrapper[10444]: I1205 10:37:40.795337 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795415 master-0 kubenswrapper[10444]: I1205 10:37:40.795362 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-cabundle\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:37:40.795415 master-0 kubenswrapper[10444]: I1205 10:37:40.795387 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795415 master-0 kubenswrapper[10444]: I1205 10:37:40.795394 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-config\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:40.795582 master-0 kubenswrapper[10444]: I1205 10:37:40.795442 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-netd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795582 master-0 kubenswrapper[10444]: I1205 10:37:40.795476 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-netd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.795582 master-0 kubenswrapper[10444]: I1205 10:37:40.795529 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:40.795722 master-0 kubenswrapper[10444]: E1205 10:37:40.795593 10444 secret.go:189] Couldn't get secret openshift-route-controller-manager/serving-cert: secret "serving-cert" not found Dec 05 10:37:40.795722 master-0 kubenswrapper[10444]: E1205 10:37:40.795624 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert podName:98fd575c-ffc4-44c2-9947-c18d3f53e2c9 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:41.295615781 +0000 UTC m=+1.887528348 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert") pod "route-controller-manager-76d4564964-xm2tr" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9") : secret "serving-cert" not found Dec 05 10:37:40.795722 master-0 kubenswrapper[10444]: I1205 10:37:40.795591 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn6sc\" (UniqueName: \"kubernetes.io/projected/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-kube-api-access-nn6sc\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:37:40.795877 master-0 kubenswrapper[10444]: I1205 10:37:40.795853 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-var-lib-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.796007 master-0 kubenswrapper[10444]: I1205 10:37:40.795898 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-systemd-units\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.796059 master-0 kubenswrapper[10444]: I1205 10:37:40.795937 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-var-lib-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.796059 master-0 kubenswrapper[10444]: I1205 10:37:40.795964 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-systemd-units\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.796136 master-0 kubenswrapper[10444]: I1205 10:37:40.796067 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/283122ba-be1c-4516-bd0f-df41c13c098b-host-slash\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:40.796176 master-0 kubenswrapper[10444]: I1205 10:37:40.796143 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-slash\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.796220 master-0 kubenswrapper[10444]: I1205 10:37:40.796185 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/283122ba-be1c-4516-bd0f-df41c13c098b-host-slash\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:40.796220 master-0 kubenswrapper[10444]: I1205 10:37:40.796213 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-slash\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.796220 master-0 kubenswrapper[10444]: I1205 10:37:40.796216 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.796381 master-0 kubenswrapper[10444]: I1205 10:37:40.796189 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:40.796999 master-0 kubenswrapper[10444]: I1205 10:37:40.796479 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-config\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:40.807429 master-0 kubenswrapper[10444]: I1205 10:37:40.807372 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 10:37:40.827812 master-0 kubenswrapper[10444]: I1205 10:37:40.827739 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 10:37:40.835812 master-0 kubenswrapper[10444]: I1205 10:37:40.835753 10444 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 05 10:37:40.841165 master-0 kubenswrapper[10444]: I1205 10:37:40.841106 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-key\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:37:40.847753 master-0 kubenswrapper[10444]: I1205 10:37:40.847701 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 10:37:40.867827 master-0 kubenswrapper[10444]: I1205 10:37:40.867756 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 10:37:40.876536 master-0 kubenswrapper[10444]: I1205 10:37:40.876471 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-cabundle\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:37:40.888066 master-0 kubenswrapper[10444]: I1205 10:37:40.888010 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 10:37:40.891368 master-0 kubenswrapper[10444]: I1205 10:37:40.891330 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/283122ba-be1c-4516-bd0f-df41c13c098b-iptables-alerter-script\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:40.926322 master-0 kubenswrapper[10444]: I1205 10:37:40.926266 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwqkb\" (UniqueName: \"kubernetes.io/projected/1a0235af-2cf2-4ad4-b419-764fb56a0107-kube-api-access-cwqkb\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:37:40.940378 master-0 kubenswrapper[10444]: I1205 10:37:40.940316 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khfxz\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-kube-api-access-khfxz\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:40.964741 master-0 kubenswrapper[10444]: I1205 10:37:40.964663 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lwgq\" (UniqueName: \"kubernetes.io/projected/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-kube-api-access-2lwgq\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:40.978123 master-0 kubenswrapper[10444]: I1205 10:37:40.978071 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11f563d5-89bb-433c-956a-6d5d2492e8f1-kube-api-access\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:37:41.019375 master-0 kubenswrapper[10444]: I1205 10:37:41.019328 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fncz7\" (UniqueName: \"kubernetes.io/projected/926263c4-ec5b-41cb-9c30-0c88f636035f-kube-api-access-fncz7\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:41.041840 master-0 kubenswrapper[10444]: I1205 10:37:41.041701 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssssf\" (UniqueName: \"kubernetes.io/projected/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-kube-api-access-ssssf\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:41.050281 master-0 kubenswrapper[10444]: I1205 10:37:41.050232 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-bound-sa-token\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:41.066606 master-0 kubenswrapper[10444]: I1205 10:37:41.066563 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dtfn\" (UniqueName: \"kubernetes.io/projected/5a1bdc70-6412-47e0-8330-04d796cc8d55-kube-api-access-5dtfn\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:41.090945 master-0 kubenswrapper[10444]: I1205 10:37:41.090872 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkmvj\" (UniqueName: \"kubernetes.io/projected/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-kube-api-access-fkmvj\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:37:41.100400 master-0 kubenswrapper[10444]: I1205 10:37:41.100339 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twh85\" (UniqueName: \"kubernetes.io/projected/f7a08359-0379-4364-8b0c-ddb58ff605f4-kube-api-access-twh85\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:37:41.122985 master-0 kubenswrapper[10444]: I1205 10:37:41.122894 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-bound-sa-token\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:41.144687 master-0 kubenswrapper[10444]: I1205 10:37:41.144616 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98qt8\" (UniqueName: \"kubernetes.io/projected/c22d947f-a5b6-4f24-b142-dd201c46293b-kube-api-access-98qt8\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:41.160849 master-0 kubenswrapper[10444]: I1205 10:37:41.160787 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfz87\" (UniqueName: \"kubernetes.io/projected/b760849c-8d83-47da-8677-68445c143bef-kube-api-access-jfz87\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:37:41.180375 master-0 kubenswrapper[10444]: I1205 10:37:41.180303 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fbss\" (UniqueName: \"kubernetes.io/projected/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-kube-api-access-2fbss\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:37:41.201523 master-0 kubenswrapper[10444]: I1205 10:37:41.201386 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:41.201723 master-0 kubenswrapper[10444]: E1205 10:37:41.201658 10444 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:41.201774 master-0 kubenswrapper[10444]: E1205 10:37:41.201737 10444 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: secret "olm-operator-serving-cert" not found Dec 05 10:37:41.201819 master-0 kubenswrapper[10444]: I1205 10:37:41.201672 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:41.201819 master-0 kubenswrapper[10444]: E1205 10:37:41.201803 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls podName:a722cda9-29a0-4b7f-8e1d-9a8950ed765a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.201761524 +0000 UTC m=+2.793674091 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-7ff994598c-kq8qr" (UID: "a722cda9-29a0-4b7f-8e1d-9a8950ed765a") : secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:41.201947 master-0 kubenswrapper[10444]: E1205 10:37:41.201825 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert podName:46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.201816915 +0000 UTC m=+2.793729572 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert") pod "olm-operator-7cd7dbb44c-d25sk" (UID: "46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082") : secret "olm-operator-serving-cert" not found Dec 05 10:37:41.201947 master-0 kubenswrapper[10444]: I1205 10:37:41.201913 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:41.202021 master-0 kubenswrapper[10444]: I1205 10:37:41.201986 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:41.202090 master-0 kubenswrapper[10444]: I1205 10:37:41.202051 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:41.202132 master-0 kubenswrapper[10444]: E1205 10:37:41.202121 10444 secret.go:189] Couldn't get secret openshift-ingress-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:41.202165 master-0 kubenswrapper[10444]: E1205 10:37:41.202127 10444 secret.go:189] Couldn't get secret openshift-image-registry/image-registry-operator-tls: secret "image-registry-operator-tls" not found Dec 05 10:37:41.202165 master-0 kubenswrapper[10444]: I1205 10:37:41.202152 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:41.202244 master-0 kubenswrapper[10444]: E1205 10:37:41.202160 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls podName:22676fac-b770-4937-9bee-7478bd1babb7 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.202148876 +0000 UTC m=+2.794061443 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls") pod "ingress-operator-8649c48786-cgt5x" (UID: "22676fac-b770-4937-9bee-7478bd1babb7") : secret "metrics-tls" not found Dec 05 10:37:41.202244 master-0 kubenswrapper[10444]: E1205 10:37:41.202213 10444 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: secret "package-server-manager-serving-cert" not found Dec 05 10:37:41.202244 master-0 kubenswrapper[10444]: E1205 10:37:41.202216 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls podName:f7b29f89-e42d-4e53-ad14-05efdce933f0 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.202205207 +0000 UTC m=+2.794117774 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls") pod "cluster-image-registry-operator-6fb9f88b7-f29mb" (UID: "f7b29f89-e42d-4e53-ad14-05efdce933f0") : secret "image-registry-operator-tls" not found Dec 05 10:37:41.202398 master-0 kubenswrapper[10444]: E1205 10:37:41.202245 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert podName:8c649a16-c187-412e-b5da-62a00bee38ab nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.202236868 +0000 UTC m=+2.794149515 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert") pod "package-server-manager-67477646d4-nm8cn" (UID: "8c649a16-c187-412e-b5da-62a00bee38ab") : secret "package-server-manager-serving-cert" not found Dec 05 10:37:41.202398 master-0 kubenswrapper[10444]: E1205 10:37:41.202294 10444 secret.go:189] Couldn't get secret openshift-dns-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:41.202398 master-0 kubenswrapper[10444]: I1205 10:37:41.202319 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:41.202398 master-0 kubenswrapper[10444]: E1205 10:37:41.202355 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls podName:5a1bdc70-6412-47e0-8330-04d796cc8d55 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.202337791 +0000 UTC m=+2.794250398 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls") pod "dns-operator-7c56cf9b74-x6t9h" (UID: "5a1bdc70-6412-47e0-8330-04d796cc8d55") : secret "metrics-tls" not found Dec 05 10:37:41.202398 master-0 kubenswrapper[10444]: E1205 10:37:41.202381 10444 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/node-tuning-operator-tls: secret "node-tuning-operator-tls" not found Dec 05 10:37:41.202626 master-0 kubenswrapper[10444]: E1205 10:37:41.202420 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.202402463 +0000 UTC m=+2.794315120 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "node-tuning-operator-tls" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "node-tuning-operator-tls" not found Dec 05 10:37:41.202626 master-0 kubenswrapper[10444]: I1205 10:37:41.202459 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:41.202626 master-0 kubenswrapper[10444]: I1205 10:37:41.202500 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:41.202626 master-0 kubenswrapper[10444]: I1205 10:37:41.202523 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:41.202626 master-0 kubenswrapper[10444]: E1205 10:37:41.202610 10444 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:37:41.202761 master-0 kubenswrapper[10444]: E1205 10:37:41.202650 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.202637929 +0000 UTC m=+2.794550496 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:37:41.202761 master-0 kubenswrapper[10444]: E1205 10:37:41.202658 10444 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: secret "multus-admission-controller-secret" not found Dec 05 10:37:41.202761 master-0 kubenswrapper[10444]: E1205 10:37:41.202721 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs podName:8d76404b-6d62-4a61-b6f6-0c8073eba198 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.202709981 +0000 UTC m=+2.794622618 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs") pod "multus-admission-controller-7dfc5b745f-67rx7" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198") : secret "multus-admission-controller-secret" not found Dec 05 10:37:41.202847 master-0 kubenswrapper[10444]: E1205 10:37:41.202755 10444 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Dec 05 10:37:41.202847 master-0 kubenswrapper[10444]: E1205 10:37:41.202798 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.202787303 +0000 UTC m=+2.794699970 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : secret "metrics-daemon-secret" not found Dec 05 10:37:41.202847 master-0 kubenswrapper[10444]: I1205 10:37:41.202838 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:41.202979 master-0 kubenswrapper[10444]: I1205 10:37:41.202941 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:41.203022 master-0 kubenswrapper[10444]: E1205 10:37:41.202982 10444 secret.go:189] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: secret "marketplace-operator-metrics" not found Dec 05 10:37:41.203022 master-0 kubenswrapper[10444]: E1205 10:37:41.203018 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics podName:926263c4-ec5b-41cb-9c30-0c88f636035f nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.203007709 +0000 UTC m=+2.794920346 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics") pod "marketplace-operator-f797b99b6-z9qcl" (UID: "926263c4-ec5b-41cb-9c30-0c88f636035f") : secret "marketplace-operator-metrics" not found Dec 05 10:37:41.203085 master-0 kubenswrapper[10444]: E1205 10:37:41.203044 10444 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/performance-addon-operator-webhook-cert: secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:41.203123 master-0 kubenswrapper[10444]: E1205 10:37:41.203089 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.203079871 +0000 UTC m=+2.794992528 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:41.206778 master-0 kubenswrapper[10444]: I1205 10:37:41.206711 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/8c649a16-c187-412e-b5da-62a00bee38ab-kube-api-access-d8pv2\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:41.223552 master-0 kubenswrapper[10444]: I1205 10:37:41.223498 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjrgm\" (UniqueName: \"kubernetes.io/projected/87909f47-f2d7-46f8-a1c8-27336cdcce5d-kube-api-access-rjrgm\") pod \"csi-snapshot-controller-operator-6bc8656fdc-vd94f\" (UID: \"87909f47-f2d7-46f8-a1c8-27336cdcce5d\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:37:41.239058 master-0 kubenswrapper[10444]: I1205 10:37:41.238992 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrdsv\" (UniqueName: \"kubernetes.io/projected/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-kube-api-access-jrdsv\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:37:41.264485 master-0 kubenswrapper[10444]: I1205 10:37:41.264408 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/444f8808-e454-4015-9e20-429e715a08c7-kube-api-access\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:37:41.277737 master-0 kubenswrapper[10444]: I1205 10:37:41.277687 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkqz7\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-kube-api-access-vkqz7\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:41.299102 master-0 kubenswrapper[10444]: I1205 10:37:41.298979 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn7rj\" (UniqueName: \"kubernetes.io/projected/eb290494-a456-4f0e-9afc-f20abab1a1bf-kube-api-access-qn7rj\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:41.303838 master-0 kubenswrapper[10444]: I1205 10:37:41.303797 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:41.304128 master-0 kubenswrapper[10444]: I1205 10:37:41.304093 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:41.304210 master-0 kubenswrapper[10444]: E1205 10:37:41.304086 10444 secret.go:189] Couldn't get secret openshift-route-controller-manager/serving-cert: secret "serving-cert" not found Dec 05 10:37:41.304256 master-0 kubenswrapper[10444]: E1205 10:37:41.304144 10444 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: configmap "client-ca" not found Dec 05 10:37:41.304304 master-0 kubenswrapper[10444]: E1205 10:37:41.304245 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert podName:98fd575c-ffc4-44c2-9947-c18d3f53e2c9 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.304230912 +0000 UTC m=+2.896143479 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert") pod "route-controller-manager-76d4564964-xm2tr" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9") : secret "serving-cert" not found Dec 05 10:37:41.304304 master-0 kubenswrapper[10444]: E1205 10:37:41.304290 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca podName:98fd575c-ffc4-44c2-9947-c18d3f53e2c9 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:42.304275523 +0000 UTC m=+2.896188100 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca") pod "route-controller-manager-76d4564964-xm2tr" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9") : configmap "client-ca" not found Dec 05 10:37:41.320013 master-0 kubenswrapper[10444]: I1205 10:37:41.319954 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vp9b\" (UniqueName: \"kubernetes.io/projected/49051e6e-5a2f-45c8-bad0-374514a91c07-kube-api-access-6vp9b\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:37:41.343338 master-0 kubenswrapper[10444]: I1205 10:37:41.343237 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvnxf\" (UniqueName: \"kubernetes.io/projected/6a9f011f-36f1-4308-a365-69425c186c7f-kube-api-access-mvnxf\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:37:41.366257 master-0 kubenswrapper[10444]: I1205 10:37:41.366190 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-kube-api-access\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:41.381105 master-0 kubenswrapper[10444]: I1205 10:37:41.381042 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4825316a-ea9f-4d3d-838b-fa809a6e49c7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:37:41.398375 master-0 kubenswrapper[10444]: I1205 10:37:41.398292 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2vd4\" (UniqueName: \"kubernetes.io/projected/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-kube-api-access-h2vd4\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:37:41.423761 master-0 kubenswrapper[10444]: I1205 10:37:41.423695 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kb5c\" (UniqueName: \"kubernetes.io/projected/8d76404b-6d62-4a61-b6f6-0c8073eba198-kube-api-access-7kb5c\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:41.440780 master-0 kubenswrapper[10444]: I1205 10:37:41.440737 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzwfq\" (UniqueName: \"kubernetes.io/projected/6f76d12f-5406-47e2-8337-2f50e35376d6-kube-api-access-fzwfq\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:37:41.446033 master-0 kubenswrapper[10444]: I1205 10:37:41.445395 10444 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:37:41.460368 master-0 kubenswrapper[10444]: I1205 10:37:41.460315 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wfsv\" (UniqueName: \"kubernetes.io/projected/1e69ce9e-4e6f-4015-9ba6-5a7942570190-kube-api-access-7wfsv\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:37:41.478719 master-0 kubenswrapper[10444]: I1205 10:37:41.478679 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv4zs\" (UniqueName: \"kubernetes.io/projected/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-kube-api-access-jv4zs\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:41.512999 master-0 kubenswrapper[10444]: E1205 10:37:41.512942 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"bootstrap-kube-scheduler-master-0\" already exists" pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:37:41.579917 master-0 kubenswrapper[10444]: W1205 10:37:41.579786 10444 warnings.go:70] would violate PodSecurity "restricted:latest": host namespaces (hostNetwork=true), hostPort (container "etcd" uses hostPorts 2379, 2380), privileged (containers "etcdctl", "etcd" must not set securityContext.privileged=true), allowPrivilegeEscalation != false (containers "etcdctl", "etcd" must set securityContext.allowPrivilegeEscalation=false), unrestricted capabilities (containers "etcdctl", "etcd" must set securityContext.capabilities.drop=["ALL"]), restricted volume types (volumes "certs", "data-dir" use restricted volume type "hostPath"), runAsNonRoot != true (pod or containers "etcdctl", "etcd" must set securityContext.runAsNonRoot=true), seccompProfile (pod or containers "etcdctl", "etcd" must set securityContext.seccompProfile.type to "RuntimeDefault" or "Localhost") Dec 05 10:37:41.579917 master-0 kubenswrapper[10444]: E1205 10:37:41.579887 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-master-0-master-0\" already exists" pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:37:41.582467 master-0 kubenswrapper[10444]: E1205 10:37:41.582416 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"bootstrap-kube-apiserver-master-0\" already exists" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:41.582822 master-0 kubenswrapper[10444]: E1205 10:37:41.582772 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"bootstrap-kube-controller-manager-master-0\" already exists" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:41.600584 master-0 kubenswrapper[10444]: I1205 10:37:41.600533 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml74w\" (UniqueName: \"kubernetes.io/projected/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f-kube-api-access-ml74w\") pod \"csi-snapshot-controller-6b958b6f94-lgn6v\" (UID: \"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" Dec 05 10:37:41.617745 master-0 kubenswrapper[10444]: I1205 10:37:41.617704 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdsh9\" (UniqueName: \"kubernetes.io/projected/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a-kube-api-access-pdsh9\") pod \"migrator-74b7b57c65-sfvzd\" (UID: \"3c4f767d-65f9-434b-8ddd-ceb0b91ab99a\") " pod="openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd" Dec 05 10:37:41.638209 master-0 kubenswrapper[10444]: I1205 10:37:41.638137 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:41.663597 master-0 kubenswrapper[10444]: I1205 10:37:41.663528 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjgc4\" (UniqueName: \"kubernetes.io/projected/495ba1ea-f844-43ec-8be7-47e738f5428a-kube-api-access-tjgc4\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:41.679450 master-0 kubenswrapper[10444]: I1205 10:37:41.679169 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhjhk\" (UniqueName: \"kubernetes.io/projected/283122ba-be1c-4516-bd0f-df41c13c098b-kube-api-access-vhjhk\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:37:41.699568 master-0 kubenswrapper[10444]: I1205 10:37:41.699494 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg9zq\" (UniqueName: \"kubernetes.io/projected/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-kube-api-access-wg9zq\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:37:41.725957 master-0 kubenswrapper[10444]: I1205 10:37:41.725836 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwpcs\" (UniqueName: \"kubernetes.io/projected/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-kube-api-access-fwpcs\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:41.743196 master-0 kubenswrapper[10444]: I1205 10:37:41.743103 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn6sc\" (UniqueName: \"kubernetes.io/projected/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-kube-api-access-nn6sc\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:37:41.841839 master-0 kubenswrapper[10444]: I1205 10:37:41.841659 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:37:42.015810 master-0 kubenswrapper[10444]: I1205 10:37:42.011184 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-77c99c46b8-m7zqs"] Dec 05 10:37:42.072063 master-0 kubenswrapper[10444]: I1205 10:37:42.072010 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:42.098027 master-0 kubenswrapper[10444]: I1205 10:37:42.097938 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:42.213685 master-0 kubenswrapper[10444]: I1205 10:37:42.213634 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:42.213879 master-0 kubenswrapper[10444]: I1205 10:37:42.213695 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:42.213879 master-0 kubenswrapper[10444]: I1205 10:37:42.213743 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:42.213986 master-0 kubenswrapper[10444]: I1205 10:37:42.213942 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:42.214037 master-0 kubenswrapper[10444]: I1205 10:37:42.214012 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:42.214099 master-0 kubenswrapper[10444]: E1205 10:37:42.214021 10444 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/performance-addon-operator-webhook-cert: secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:42.214134 master-0 kubenswrapper[10444]: E1205 10:37:42.214112 10444 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:42.214166 master-0 kubenswrapper[10444]: E1205 10:37:42.214134 10444 secret.go:189] Couldn't get secret openshift-image-registry/image-registry-operator-tls: secret "image-registry-operator-tls" not found Dec 05 10:37:42.214166 master-0 kubenswrapper[10444]: I1205 10:37:42.214063 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:42.214166 master-0 kubenswrapper[10444]: E1205 10:37:42.214139 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.214119035 +0000 UTC m=+4.806031602 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "performance-addon-operator-webhook-cert" not found Dec 05 10:37:42.214250 master-0 kubenswrapper[10444]: E1205 10:37:42.214087 10444 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: secret "olm-operator-serving-cert" not found Dec 05 10:37:42.214250 master-0 kubenswrapper[10444]: E1205 10:37:42.214191 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls podName:a722cda9-29a0-4b7f-8e1d-9a8950ed765a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.214173167 +0000 UTC m=+4.806085784 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-7ff994598c-kq8qr" (UID: "a722cda9-29a0-4b7f-8e1d-9a8950ed765a") : secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:42.214250 master-0 kubenswrapper[10444]: E1205 10:37:42.214055 10444 secret.go:189] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: secret "marketplace-operator-metrics" not found Dec 05 10:37:42.214250 master-0 kubenswrapper[10444]: E1205 10:37:42.214205 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls podName:f7b29f89-e42d-4e53-ad14-05efdce933f0 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.214198748 +0000 UTC m=+4.806111385 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls") pod "cluster-image-registry-operator-6fb9f88b7-f29mb" (UID: "f7b29f89-e42d-4e53-ad14-05efdce933f0") : secret "image-registry-operator-tls" not found Dec 05 10:37:42.214250 master-0 kubenswrapper[10444]: I1205 10:37:42.214223 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:42.214416 master-0 kubenswrapper[10444]: E1205 10:37:42.214253 10444 secret.go:189] Couldn't get secret openshift-ingress-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:42.214416 master-0 kubenswrapper[10444]: I1205 10:37:42.214259 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:42.214416 master-0 kubenswrapper[10444]: E1205 10:37:42.214280 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls podName:22676fac-b770-4937-9bee-7478bd1babb7 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.21426914 +0000 UTC m=+4.806181757 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls") pod "ingress-operator-8649c48786-cgt5x" (UID: "22676fac-b770-4937-9bee-7478bd1babb7") : secret "metrics-tls" not found Dec 05 10:37:42.214416 master-0 kubenswrapper[10444]: I1205 10:37:42.214343 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:42.214416 master-0 kubenswrapper[10444]: E1205 10:37:42.214355 10444 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: secret "package-server-manager-serving-cert" not found Dec 05 10:37:42.214416 master-0 kubenswrapper[10444]: I1205 10:37:42.214376 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:42.214416 master-0 kubenswrapper[10444]: E1205 10:37:42.214383 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert podName:8c649a16-c187-412e-b5da-62a00bee38ab nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.214375023 +0000 UTC m=+4.806287640 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert") pod "package-server-manager-67477646d4-nm8cn" (UID: "8c649a16-c187-412e-b5da-62a00bee38ab") : secret "package-server-manager-serving-cert" not found Dec 05 10:37:42.214416 master-0 kubenswrapper[10444]: E1205 10:37:42.214408 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert podName:46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.214401384 +0000 UTC m=+4.806313951 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert") pod "olm-operator-7cd7dbb44c-d25sk" (UID: "46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082") : secret "olm-operator-serving-cert" not found Dec 05 10:37:42.214416 master-0 kubenswrapper[10444]: E1205 10:37:42.214420 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics podName:926263c4-ec5b-41cb-9c30-0c88f636035f nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.214413864 +0000 UTC m=+4.806326431 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics") pod "marketplace-operator-f797b99b6-z9qcl" (UID: "926263c4-ec5b-41cb-9c30-0c88f636035f") : secret "marketplace-operator-metrics" not found Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: E1205 10:37:42.214444 10444 secret.go:189] Couldn't get secret openshift-dns-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: I1205 10:37:42.214448 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: E1205 10:37:42.214470 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls podName:5a1bdc70-6412-47e0-8330-04d796cc8d55 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.214462295 +0000 UTC m=+4.806374862 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls") pod "dns-operator-7c56cf9b74-x6t9h" (UID: "5a1bdc70-6412-47e0-8330-04d796cc8d55") : secret "metrics-tls" not found Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: I1205 10:37:42.214485 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: E1205 10:37:42.214496 10444 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: secret "multus-admission-controller-secret" not found Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: E1205 10:37:42.214515 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs podName:8d76404b-6d62-4a61-b6f6-0c8073eba198 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.214509306 +0000 UTC m=+4.806421873 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs") pod "multus-admission-controller-7dfc5b745f-67rx7" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198") : secret "multus-admission-controller-secret" not found Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: E1205 10:37:42.213946 10444 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: E1205 10:37:42.214564 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.214555968 +0000 UTC m=+4.806468535 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : secret "metrics-daemon-secret" not found Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: E1205 10:37:42.214564 10444 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: E1205 10:37:42.214587 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.214580598 +0000 UTC m=+4.806493235 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: E1205 10:37:42.214596 10444 secret.go:189] Couldn't get secret openshift-cluster-node-tuning-operator/node-tuning-operator-tls: secret "node-tuning-operator-tls" not found Dec 05 10:37:42.214677 master-0 kubenswrapper[10444]: E1205 10:37:42.214634 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls podName:eb290494-a456-4f0e-9afc-f20abab1a1bf nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.21462402 +0000 UTC m=+4.806536627 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "node-tuning-operator-tls" (UniqueName: "kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls") pod "cluster-node-tuning-operator-85cff47f46-qwx2p" (UID: "eb290494-a456-4f0e-9afc-f20abab1a1bf") : secret "node-tuning-operator-tls" not found Dec 05 10:37:42.314936 master-0 kubenswrapper[10444]: I1205 10:37:42.314881 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:42.315114 master-0 kubenswrapper[10444]: I1205 10:37:42.314975 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:42.315114 master-0 kubenswrapper[10444]: E1205 10:37:42.315085 10444 secret.go:189] Couldn't get secret openshift-route-controller-manager/serving-cert: secret "serving-cert" not found Dec 05 10:37:42.315194 master-0 kubenswrapper[10444]: E1205 10:37:42.315127 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert podName:98fd575c-ffc4-44c2-9947-c18d3f53e2c9 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.315111961 +0000 UTC m=+4.907024528 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert") pod "route-controller-manager-76d4564964-xm2tr" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9") : secret "serving-cert" not found Dec 05 10:37:42.315504 master-0 kubenswrapper[10444]: E1205 10:37:42.315412 10444 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: configmap "client-ca" not found Dec 05 10:37:42.315504 master-0 kubenswrapper[10444]: E1205 10:37:42.315465 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca podName:98fd575c-ffc4-44c2-9947-c18d3f53e2c9 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.31545639 +0000 UTC m=+4.907368957 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca") pod "route-controller-manager-76d4564964-xm2tr" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9") : configmap "client-ca" not found Dec 05 10:37:42.494611 master-0 kubenswrapper[10444]: I1205 10:37:42.494470 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-77f4fc6d5d-zdn92"] Dec 05 10:37:42.497807 master-0 kubenswrapper[10444]: I1205 10:37:42.497764 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-77f4fc6d5d-zdn92"] Dec 05 10:37:42.698349 master-0 kubenswrapper[10444]: I1205 10:37:42.698117 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" event={"ID":"97192d4d-15d3-4740-82f3-d0d45f9fe7b9","Type":"ContainerStarted","Data":"1d5d08b7e3a6434c78fb95ac9b4c4e5bd43fdee37f382d3c530fa68d7a00cb3f"} Dec 05 10:37:42.698349 master-0 kubenswrapper[10444]: I1205 10:37:42.698184 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" event={"ID":"97192d4d-15d3-4740-82f3-d0d45f9fe7b9","Type":"ContainerStarted","Data":"088f93fb71215e03078ad676d01b6c7b68c855017eff77ac0edbcdeaa645fd6c"} Dec 05 10:37:42.703577 master-0 kubenswrapper[10444]: I1205 10:37:42.703473 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-d6fzk" event={"ID":"720a7467-ce93-4d48-82ec-9ad0922d99c2","Type":"ContainerStarted","Data":"594924be7583d6c7cadf5cbf1494e6c57995d5c1c4e1725f6d24bfecd89d204a"} Dec 05 10:37:42.838308 master-0 kubenswrapper[10444]: I1205 10:37:42.838229 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" podStartSLOduration=3.8382066889999997 podStartE2EDuration="3.838206689s" podCreationTimestamp="2025-12-05 10:37:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:37:42.836398638 +0000 UTC m=+3.428311205" watchObservedRunningTime="2025-12-05 10:37:42.838206689 +0000 UTC m=+3.430119256" Dec 05 10:37:43.361456 master-0 kubenswrapper[10444]: I1205 10:37:43.360882 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6"] Dec 05 10:37:43.361456 master-0 kubenswrapper[10444]: E1205 10:37:43.361039 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af980e3-b7b6-4a5b-96c4-53cbbe43c268" containerName="prober" Dec 05 10:37:43.361456 master-0 kubenswrapper[10444]: I1205 10:37:43.361050 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af980e3-b7b6-4a5b-96c4-53cbbe43c268" containerName="prober" Dec 05 10:37:43.361456 master-0 kubenswrapper[10444]: E1205 10:37:43.361060 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1d74f88-1419-431e-80da-26db419f050e" containerName="assisted-installer-controller" Dec 05 10:37:43.361456 master-0 kubenswrapper[10444]: I1205 10:37:43.361066 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1d74f88-1419-431e-80da-26db419f050e" containerName="assisted-installer-controller" Dec 05 10:37:43.361456 master-0 kubenswrapper[10444]: I1205 10:37:43.361120 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="3af980e3-b7b6-4a5b-96c4-53cbbe43c268" containerName="prober" Dec 05 10:37:43.361456 master-0 kubenswrapper[10444]: I1205 10:37:43.361131 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1d74f88-1419-431e-80da-26db419f050e" containerName="assisted-installer-controller" Dec 05 10:37:43.362648 master-0 kubenswrapper[10444]: I1205 10:37:43.361897 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.365308 master-0 kubenswrapper[10444]: I1205 10:37:43.365250 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 10:37:43.366875 master-0 kubenswrapper[10444]: I1205 10:37:43.366321 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 10:37:43.366875 master-0 kubenswrapper[10444]: I1205 10:37:43.366574 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 10:37:43.366875 master-0 kubenswrapper[10444]: I1205 10:37:43.366636 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 10:37:43.366875 master-0 kubenswrapper[10444]: I1205 10:37:43.366646 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 10:37:43.375157 master-0 kubenswrapper[10444]: I1205 10:37:43.375112 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 10:37:43.414466 master-0 kubenswrapper[10444]: I1205 10:37:43.413567 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6"] Dec 05 10:37:43.429458 master-0 kubenswrapper[10444]: I1205 10:37:43.427190 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfh6j\" (UniqueName: \"kubernetes.io/projected/61fd412b-c1c2-4902-9564-aab7a8b17845-kube-api-access-tfh6j\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.429458 master-0 kubenswrapper[10444]: I1205 10:37:43.427262 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-proxy-ca-bundles\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.429458 master-0 kubenswrapper[10444]: I1205 10:37:43.427327 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.429458 master-0 kubenswrapper[10444]: I1205 10:37:43.427359 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-config\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.429458 master-0 kubenswrapper[10444]: I1205 10:37:43.427459 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.529191 master-0 kubenswrapper[10444]: I1205 10:37:43.528457 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.529191 master-0 kubenswrapper[10444]: I1205 10:37:43.528548 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfh6j\" (UniqueName: \"kubernetes.io/projected/61fd412b-c1c2-4902-9564-aab7a8b17845-kube-api-access-tfh6j\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.529191 master-0 kubenswrapper[10444]: I1205 10:37:43.528586 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-proxy-ca-bundles\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.529191 master-0 kubenswrapper[10444]: I1205 10:37:43.528622 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.529191 master-0 kubenswrapper[10444]: I1205 10:37:43.528668 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-config\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.529754 master-0 kubenswrapper[10444]: E1205 10:37:43.529532 10444 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: configmap "client-ca" not found Dec 05 10:37:43.529754 master-0 kubenswrapper[10444]: E1205 10:37:43.529583 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca podName:61fd412b-c1c2-4902-9564-aab7a8b17845 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.029564165 +0000 UTC m=+4.621476742 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca") pod "controller-manager-6d9cb7b7fc-f9nz6" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845") : configmap "client-ca" not found Dec 05 10:37:43.530082 master-0 kubenswrapper[10444]: E1205 10:37:43.529975 10444 secret.go:189] Couldn't get secret openshift-controller-manager/serving-cert: secret "serving-cert" not found Dec 05 10:37:43.530082 master-0 kubenswrapper[10444]: E1205 10:37:43.530059 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert podName:61fd412b-c1c2-4902-9564-aab7a8b17845 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:44.030038138 +0000 UTC m=+4.621950705 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert") pod "controller-manager-6d9cb7b7fc-f9nz6" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845") : secret "serving-cert" not found Dec 05 10:37:43.532480 master-0 kubenswrapper[10444]: I1205 10:37:43.530818 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-proxy-ca-bundles\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.532480 master-0 kubenswrapper[10444]: I1205 10:37:43.531448 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-config\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:43.706713 master-0 kubenswrapper[10444]: I1205 10:37:43.706573 10444 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:37:43.707238 master-0 kubenswrapper[10444]: I1205 10:37:43.707226 10444 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:37:44.034829 master-0 kubenswrapper[10444]: I1205 10:37:44.034759 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:44.034999 master-0 kubenswrapper[10444]: I1205 10:37:44.034923 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:44.035255 master-0 kubenswrapper[10444]: E1205 10:37:44.035128 10444 secret.go:189] Couldn't get secret openshift-controller-manager/serving-cert: secret "serving-cert" not found Dec 05 10:37:44.035255 master-0 kubenswrapper[10444]: E1205 10:37:44.035173 10444 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: configmap "client-ca" not found Dec 05 10:37:44.035255 master-0 kubenswrapper[10444]: E1205 10:37:44.035230 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert podName:61fd412b-c1c2-4902-9564-aab7a8b17845 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:45.0352043 +0000 UTC m=+5.627116887 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert") pod "controller-manager-6d9cb7b7fc-f9nz6" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845") : secret "serving-cert" not found Dec 05 10:37:44.035395 master-0 kubenswrapper[10444]: E1205 10:37:44.035258 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca podName:61fd412b-c1c2-4902-9564-aab7a8b17845 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:45.035245401 +0000 UTC m=+5.627157978 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca") pod "controller-manager-6d9cb7b7fc-f9nz6" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845") : configmap "client-ca" not found Dec 05 10:37:44.092349 master-0 kubenswrapper[10444]: I1205 10:37:44.092083 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfh6j\" (UniqueName: \"kubernetes.io/projected/61fd412b-c1c2-4902-9564-aab7a8b17845-kube-api-access-tfh6j\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:44.199255 master-0 kubenswrapper[10444]: I1205 10:37:44.198618 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.236822 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.236883 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.236905 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.236930 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.236953 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.236974 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.237016 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.237034 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.237059 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.237080 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.237118 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: I1205 10:37:44.237147 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237285 10444 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237362 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls podName:a722cda9-29a0-4b7f-8e1d-9a8950ed765a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.237342034 +0000 UTC m=+8.829254601 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-7ff994598c-kq8qr" (UID: "a722cda9-29a0-4b7f-8e1d-9a8950ed765a") : secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237466 10444 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: secret "olm-operator-serving-cert" not found Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237501 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert podName:46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.23749105 +0000 UTC m=+8.829403617 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert") pod "olm-operator-7cd7dbb44c-d25sk" (UID: "46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082") : secret "olm-operator-serving-cert" not found Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237553 10444 secret.go:189] Couldn't get secret openshift-image-registry/image-registry-operator-tls: secret "image-registry-operator-tls" not found Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237582 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls podName:f7b29f89-e42d-4e53-ad14-05efdce933f0 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.237573552 +0000 UTC m=+8.829486119 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "image-registry-operator-tls" (UniqueName: "kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls") pod "cluster-image-registry-operator-6fb9f88b7-f29mb" (UID: "f7b29f89-e42d-4e53-ad14-05efdce933f0") : secret "image-registry-operator-tls" not found Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237628 10444 secret.go:189] Couldn't get secret openshift-ingress-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237653 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls podName:22676fac-b770-4937-9bee-7478bd1babb7 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.237644794 +0000 UTC m=+8.829557361 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls") pod "ingress-operator-8649c48786-cgt5x" (UID: "22676fac-b770-4937-9bee-7478bd1babb7") : secret "metrics-tls" not found Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237711 10444 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: secret "package-server-manager-serving-cert" not found Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237736 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert podName:8c649a16-c187-412e-b5da-62a00bee38ab nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.237728716 +0000 UTC m=+8.829641283 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert") pod "package-server-manager-67477646d4-nm8cn" (UID: "8c649a16-c187-412e-b5da-62a00bee38ab") : secret "package-server-manager-serving-cert" not found Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237777 10444 secret.go:189] Couldn't get secret openshift-dns-operator/metrics-tls: secret "metrics-tls" not found Dec 05 10:37:44.238049 master-0 kubenswrapper[10444]: E1205 10:37:44.237799 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls podName:5a1bdc70-6412-47e0-8330-04d796cc8d55 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.237791768 +0000 UTC m=+8.829704345 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls") pod "dns-operator-7c56cf9b74-x6t9h" (UID: "5a1bdc70-6412-47e0-8330-04d796cc8d55") : secret "metrics-tls" not found Dec 05 10:37:44.238934 master-0 kubenswrapper[10444]: E1205 10:37:44.238413 10444 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Dec 05 10:37:44.238934 master-0 kubenswrapper[10444]: E1205 10:37:44.238467 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.238456096 +0000 UTC m=+8.830368663 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : secret "metrics-daemon-secret" not found Dec 05 10:37:44.238934 master-0 kubenswrapper[10444]: E1205 10:37:44.238543 10444 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: secret "multus-admission-controller-secret" not found Dec 05 10:37:44.238934 master-0 kubenswrapper[10444]: E1205 10:37:44.238571 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs podName:8d76404b-6d62-4a61-b6f6-0c8073eba198 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.238562719 +0000 UTC m=+8.830475286 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs") pod "multus-admission-controller-7dfc5b745f-67rx7" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198") : secret "multus-admission-controller-secret" not found Dec 05 10:37:44.238934 master-0 kubenswrapper[10444]: E1205 10:37:44.238616 10444 secret.go:189] Couldn't get secret openshift-cluster-version/cluster-version-operator-serving-cert: secret "cluster-version-operator-serving-cert" not found Dec 05 10:37:44.238934 master-0 kubenswrapper[10444]: E1205 10:37:44.238640 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert podName:82ef99d4-41b7-4859-a5a1-4e31ce614a2a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.238633011 +0000 UTC m=+8.830545578 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert") pod "cluster-version-operator-77dfcc565f-bv84m" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a") : secret "cluster-version-operator-serving-cert" not found Dec 05 10:37:44.239536 master-0 kubenswrapper[10444]: E1205 10:37:44.239323 10444 secret.go:189] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: secret "marketplace-operator-metrics" not found Dec 05 10:37:44.239536 master-0 kubenswrapper[10444]: E1205 10:37:44.239491 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics podName:926263c4-ec5b-41cb-9c30-0c88f636035f nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.239411183 +0000 UTC m=+8.831323780 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics") pod "marketplace-operator-f797b99b6-z9qcl" (UID: "926263c4-ec5b-41cb-9c30-0c88f636035f") : secret "marketplace-operator-metrics" not found Dec 05 10:37:44.243444 master-0 kubenswrapper[10444]: I1205 10:37:44.241716 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:44.258162 master-0 kubenswrapper[10444]: I1205 10:37:44.258109 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:44.337940 master-0 kubenswrapper[10444]: I1205 10:37:44.337823 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:44.337940 master-0 kubenswrapper[10444]: I1205 10:37:44.337925 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:44.338128 master-0 kubenswrapper[10444]: E1205 10:37:44.338041 10444 secret.go:189] Couldn't get secret openshift-route-controller-manager/serving-cert: secret "serving-cert" not found Dec 05 10:37:44.338128 master-0 kubenswrapper[10444]: E1205 10:37:44.338088 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert podName:98fd575c-ffc4-44c2-9947-c18d3f53e2c9 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.338073832 +0000 UTC m=+8.929986399 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert") pod "route-controller-manager-76d4564964-xm2tr" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9") : secret "serving-cert" not found Dec 05 10:37:44.338504 master-0 kubenswrapper[10444]: E1205 10:37:44.338479 10444 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: configmap "client-ca" not found Dec 05 10:37:44.338588 master-0 kubenswrapper[10444]: E1205 10:37:44.338517 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca podName:98fd575c-ffc4-44c2-9947-c18d3f53e2c9 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:48.338508615 +0000 UTC m=+8.930421182 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca") pod "route-controller-manager-76d4564964-xm2tr" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9") : configmap "client-ca" not found Dec 05 10:37:44.442105 master-0 kubenswrapper[10444]: I1205 10:37:44.442025 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:37:44.779172 master-0 kubenswrapper[10444]: I1205 10:37:44.778699 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p"] Dec 05 10:37:44.803127 master-0 kubenswrapper[10444]: W1205 10:37:44.803045 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb290494_a456_4f0e_9afc_f20abab1a1bf.slice/crio-97c0ac3e4d477fff0dd9953d178193fc87456dca36efdf5e88346f7b00aeff5b WatchSource:0}: Error finding container 97c0ac3e4d477fff0dd9953d178193fc87456dca36efdf5e88346f7b00aeff5b: Status 404 returned error can't find the container with id 97c0ac3e4d477fff0dd9953d178193fc87456dca36efdf5e88346f7b00aeff5b Dec 05 10:37:44.865685 master-0 kubenswrapper[10444]: I1205 10:37:44.865337 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:44.873034 master-0 kubenswrapper[10444]: I1205 10:37:44.872857 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:45.045273 master-0 kubenswrapper[10444]: I1205 10:37:45.044608 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:45.048762 master-0 kubenswrapper[10444]: I1205 10:37:45.045875 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:45.048762 master-0 kubenswrapper[10444]: E1205 10:37:45.046116 10444 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: configmap "client-ca" not found Dec 05 10:37:45.048762 master-0 kubenswrapper[10444]: E1205 10:37:45.046182 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca podName:61fd412b-c1c2-4902-9564-aab7a8b17845 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:47.046163051 +0000 UTC m=+7.638075618 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca") pod "controller-manager-6d9cb7b7fc-f9nz6" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845") : configmap "client-ca" not found Dec 05 10:37:45.050519 master-0 kubenswrapper[10444]: I1205 10:37:45.050478 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:45.721671 master-0 kubenswrapper[10444]: I1205 10:37:45.721522 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" event={"ID":"c22d947f-a5b6-4f24-b142-dd201c46293b","Type":"ContainerStarted","Data":"3e549526f5057891751a60c691572ec799806d7061304478b0c1e902a36e6208"} Dec 05 10:37:45.722326 master-0 kubenswrapper[10444]: I1205 10:37:45.722100 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:45.723067 master-0 kubenswrapper[10444]: I1205 10:37:45.723005 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" event={"ID":"eb290494-a456-4f0e-9afc-f20abab1a1bf","Type":"ContainerStarted","Data":"97c0ac3e4d477fff0dd9953d178193fc87456dca36efdf5e88346f7b00aeff5b"} Dec 05 10:37:45.725602 master-0 kubenswrapper[10444]: I1205 10:37:45.725538 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd" event={"ID":"3c4f767d-65f9-434b-8ddd-ceb0b91ab99a","Type":"ContainerStarted","Data":"53c93b112732c110c6d7544b3dbf4c38003771841d8c4201db004482c42633ce"} Dec 05 10:37:45.725602 master-0 kubenswrapper[10444]: I1205 10:37:45.725596 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd" event={"ID":"3c4f767d-65f9-434b-8ddd-ceb0b91ab99a","Type":"ContainerStarted","Data":"415b0a7833abdce60bc334bb2b2d1fd06895503322fc9d591bdbc9d8c0cce7d9"} Dec 05 10:37:45.731061 master-0 kubenswrapper[10444]: I1205 10:37:45.731003 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:37:46.219327 master-0 kubenswrapper[10444]: I1205 10:37:46.219238 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:46.225083 master-0 kubenswrapper[10444]: I1205 10:37:46.224941 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:47.014869 master-0 kubenswrapper[10444]: I1205 10:37:47.014772 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:47.015126 master-0 kubenswrapper[10444]: I1205 10:37:47.015074 10444 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:37:47.015126 master-0 kubenswrapper[10444]: I1205 10:37:47.015097 10444 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:37:47.041391 master-0 kubenswrapper[10444]: I1205 10:37:47.041342 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:47.067679 master-0 kubenswrapper[10444]: I1205 10:37:47.067582 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:47.067924 master-0 kubenswrapper[10444]: E1205 10:37:47.067750 10444 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: configmap "client-ca" not found Dec 05 10:37:47.067924 master-0 kubenswrapper[10444]: E1205 10:37:47.067849 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca podName:61fd412b-c1c2-4902-9564-aab7a8b17845 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:51.067826397 +0000 UTC m=+11.659738964 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca") pod "controller-manager-6d9cb7b7fc-f9nz6" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845") : configmap "client-ca" not found Dec 05 10:37:47.538448 master-0 kubenswrapper[10444]: I1205 10:37:47.537762 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:47.549816 master-0 kubenswrapper[10444]: I1205 10:37:47.549637 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:47.726546 master-0 kubenswrapper[10444]: I1205 10:37:47.719577 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:47.737646 master-0 kubenswrapper[10444]: I1205 10:37:47.737604 10444 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:37:47.751495 master-0 kubenswrapper[10444]: I1205 10:37:47.749393 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:37:47.791759 master-0 kubenswrapper[10444]: I1205 10:37:47.791659 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:48.066780 master-0 kubenswrapper[10444]: I1205 10:37:48.066664 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:48.070761 master-0 kubenswrapper[10444]: I1205 10:37:48.070726 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:48.284602 master-0 kubenswrapper[10444]: I1205 10:37:48.284550 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:48.284805 master-0 kubenswrapper[10444]: I1205 10:37:48.284650 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:48.284805 master-0 kubenswrapper[10444]: I1205 10:37:48.284713 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:48.284805 master-0 kubenswrapper[10444]: I1205 10:37:48.284747 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:48.284955 master-0 kubenswrapper[10444]: I1205 10:37:48.284810 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:48.284955 master-0 kubenswrapper[10444]: I1205 10:37:48.284874 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:48.284955 master-0 kubenswrapper[10444]: I1205 10:37:48.284910 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:48.285080 master-0 kubenswrapper[10444]: I1205 10:37:48.284964 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:48.285080 master-0 kubenswrapper[10444]: I1205 10:37:48.284996 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:48.285080 master-0 kubenswrapper[10444]: I1205 10:37:48.285020 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:48.285192 master-0 kubenswrapper[10444]: E1205 10:37:48.285150 10444 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: secret "metrics-daemon-secret" not found Dec 05 10:37:48.285277 master-0 kubenswrapper[10444]: E1205 10:37:48.285206 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs podName:3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c nodeName:}" failed. No retries permitted until 2025-12-05 10:37:56.285187925 +0000 UTC m=+16.877100492 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs") pod "network-metrics-daemon-8gjgm" (UID: "3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c") : secret "metrics-daemon-secret" not found Dec 05 10:37:48.286154 master-0 kubenswrapper[10444]: E1205 10:37:48.285681 10444 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/package-server-manager-serving-cert: secret "package-server-manager-serving-cert" not found Dec 05 10:37:48.286154 master-0 kubenswrapper[10444]: E1205 10:37:48.285722 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert podName:8c649a16-c187-412e-b5da-62a00bee38ab nodeName:}" failed. No retries permitted until 2025-12-05 10:37:56.285710959 +0000 UTC m=+16.877623526 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "package-server-manager-serving-cert" (UniqueName: "kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert") pod "package-server-manager-67477646d4-nm8cn" (UID: "8c649a16-c187-412e-b5da-62a00bee38ab") : secret "package-server-manager-serving-cert" not found Dec 05 10:37:48.286287 master-0 kubenswrapper[10444]: E1205 10:37:48.286276 10444 secret.go:189] Couldn't get secret openshift-monitoring/cluster-monitoring-operator-tls: secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:48.286339 master-0 kubenswrapper[10444]: E1205 10:37:48.286315 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls podName:a722cda9-29a0-4b7f-8e1d-9a8950ed765a nodeName:}" failed. No retries permitted until 2025-12-05 10:37:56.286303995 +0000 UTC m=+16.878216562 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cluster-monitoring-operator-tls" (UniqueName: "kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls") pod "cluster-monitoring-operator-7ff994598c-kq8qr" (UID: "a722cda9-29a0-4b7f-8e1d-9a8950ed765a") : secret "cluster-monitoring-operator-tls" not found Dec 05 10:37:48.286339 master-0 kubenswrapper[10444]: E1205 10:37:48.286328 10444 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: secret "multus-admission-controller-secret" not found Dec 05 10:37:48.286467 master-0 kubenswrapper[10444]: E1205 10:37:48.286370 10444 secret.go:189] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: secret "marketplace-operator-metrics" not found Dec 05 10:37:48.286530 master-0 kubenswrapper[10444]: E1205 10:37:48.286488 10444 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: secret "olm-operator-serving-cert" not found Dec 05 10:37:48.286666 master-0 kubenswrapper[10444]: E1205 10:37:48.286374 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs podName:8d76404b-6d62-4a61-b6f6-0c8073eba198 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:56.286361107 +0000 UTC m=+16.878273674 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs") pod "multus-admission-controller-7dfc5b745f-67rx7" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198") : secret "multus-admission-controller-secret" not found Dec 05 10:37:48.286800 master-0 kubenswrapper[10444]: E1205 10:37:48.286778 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert podName:46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:56.286662706 +0000 UTC m=+16.878575273 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert") pod "olm-operator-7cd7dbb44c-d25sk" (UID: "46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082") : secret "olm-operator-serving-cert" not found Dec 05 10:37:48.286876 master-0 kubenswrapper[10444]: E1205 10:37:48.286808 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics podName:926263c4-ec5b-41cb-9c30-0c88f636035f nodeName:}" failed. No retries permitted until 2025-12-05 10:37:56.28679746 +0000 UTC m=+16.878710027 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics") pod "marketplace-operator-f797b99b6-z9qcl" (UID: "926263c4-ec5b-41cb-9c30-0c88f636035f") : secret "marketplace-operator-metrics" not found Dec 05 10:37:48.290448 master-0 kubenswrapper[10444]: I1205 10:37:48.290356 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:48.291050 master-0 kubenswrapper[10444]: I1205 10:37:48.291014 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:48.291360 master-0 kubenswrapper[10444]: I1205 10:37:48.291333 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"cluster-version-operator-77dfcc565f-bv84m\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:48.292403 master-0 kubenswrapper[10444]: I1205 10:37:48.292368 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:48.341983 master-0 kubenswrapper[10444]: I1205 10:37:48.341876 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:37:48.343939 master-0 kubenswrapper[10444]: I1205 10:37:48.343660 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:37:48.351047 master-0 kubenswrapper[10444]: I1205 10:37:48.350098 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:37:48.360902 master-0 kubenswrapper[10444]: I1205 10:37:48.360809 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:37:48.386367 master-0 kubenswrapper[10444]: I1205 10:37:48.386298 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:48.386367 master-0 kubenswrapper[10444]: I1205 10:37:48.386379 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:48.386628 master-0 kubenswrapper[10444]: E1205 10:37:48.386489 10444 secret.go:189] Couldn't get secret openshift-route-controller-manager/serving-cert: secret "serving-cert" not found Dec 05 10:37:48.386628 master-0 kubenswrapper[10444]: E1205 10:37:48.386551 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert podName:98fd575c-ffc4-44c2-9947-c18d3f53e2c9 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:56.386531949 +0000 UTC m=+16.978444516 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert") pod "route-controller-manager-76d4564964-xm2tr" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9") : secret "serving-cert" not found Dec 05 10:37:48.386688 master-0 kubenswrapper[10444]: E1205 10:37:48.386608 10444 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: configmap "client-ca" not found Dec 05 10:37:48.386717 master-0 kubenswrapper[10444]: E1205 10:37:48.386692 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca podName:98fd575c-ffc4-44c2-9947-c18d3f53e2c9 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:56.386670963 +0000 UTC m=+16.978583530 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca") pod "route-controller-manager-76d4564964-xm2tr" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9") : configmap "client-ca" not found Dec 05 10:37:48.745870 master-0 kubenswrapper[10444]: I1205 10:37:48.745740 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:37:49.801813 master-0 kubenswrapper[10444]: I1205 10:37:49.800540 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-7c8487d4d9-hsrsh"] Dec 05 10:37:49.801813 master-0 kubenswrapper[10444]: I1205 10:37:49.801794 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:49.804517 master-0 kubenswrapper[10444]: I1205 10:37:49.803751 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-0" Dec 05 10:37:49.804517 master-0 kubenswrapper[10444]: I1205 10:37:49.803939 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 10:37:49.805201 master-0 kubenswrapper[10444]: I1205 10:37:49.805163 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 10:37:49.806646 master-0 kubenswrapper[10444]: I1205 10:37:49.805376 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-0" Dec 05 10:37:49.806646 master-0 kubenswrapper[10444]: I1205 10:37:49.805567 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 10:37:49.806646 master-0 kubenswrapper[10444]: I1205 10:37:49.805747 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 10:37:49.806646 master-0 kubenswrapper[10444]: I1205 10:37:49.806411 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 10:37:49.806822 master-0 kubenswrapper[10444]: I1205 10:37:49.806703 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 10:37:49.806822 master-0 kubenswrapper[10444]: I1205 10:37:49.806731 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 10:37:49.819058 master-0 kubenswrapper[10444]: I1205 10:37:49.819005 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-7c8487d4d9-hsrsh"] Dec 05 10:37:49.825411 master-0 kubenswrapper[10444]: I1205 10:37:49.825344 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 10:37:49.906001 master-0 kubenswrapper[10444]: I1205 10:37:49.905901 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-config\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:49.906001 master-0 kubenswrapper[10444]: I1205 10:37:49.905989 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-client\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:49.906287 master-0 kubenswrapper[10444]: I1205 10:37:49.906032 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-encryption-config\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:49.906287 master-0 kubenswrapper[10444]: I1205 10:37:49.906056 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-audit-dir\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:49.906287 master-0 kubenswrapper[10444]: I1205 10:37:49.906081 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-node-pullsecrets\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:49.906287 master-0 kubenswrapper[10444]: I1205 10:37:49.906123 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-serving-ca\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:49.906287 master-0 kubenswrapper[10444]: I1205 10:37:49.906145 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:49.906287 master-0 kubenswrapper[10444]: I1205 10:37:49.906190 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:49.906287 master-0 kubenswrapper[10444]: I1205 10:37:49.906215 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-image-import-ca\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:49.906287 master-0 kubenswrapper[10444]: I1205 10:37:49.906230 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-trusted-ca-bundle\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:49.906287 master-0 kubenswrapper[10444]: I1205 10:37:49.906250 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg9f9\" (UniqueName: \"kubernetes.io/projected/430249db-e9a3-4fdb-9090-96443b92a91b-kube-api-access-wg9f9\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.006796 master-0 kubenswrapper[10444]: I1205 10:37:50.006749 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-serving-ca\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.007086 master-0 kubenswrapper[10444]: I1205 10:37:50.007025 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.007193 master-0 kubenswrapper[10444]: E1205 10:37:50.007173 10444 secret.go:189] Couldn't get secret openshift-apiserver/serving-cert: secret "serving-cert" not found Dec 05 10:37:50.007238 master-0 kubenswrapper[10444]: I1205 10:37:50.007194 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.007287 master-0 kubenswrapper[10444]: E1205 10:37:50.007239 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert podName:430249db-e9a3-4fdb-9090-96443b92a91b nodeName:}" failed. No retries permitted until 2025-12-05 10:37:50.507218328 +0000 UTC m=+11.099130895 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert") pod "apiserver-7c8487d4d9-hsrsh" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b") : secret "serving-cert" not found Dec 05 10:37:50.007287 master-0 kubenswrapper[10444]: I1205 10:37:50.007274 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-trusted-ca-bundle\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.007356 master-0 kubenswrapper[10444]: I1205 10:37:50.007300 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-image-import-ca\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.007356 master-0 kubenswrapper[10444]: E1205 10:37:50.007313 10444 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-0: configmap "audit-0" not found Dec 05 10:37:50.007356 master-0 kubenswrapper[10444]: I1205 10:37:50.007325 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg9f9\" (UniqueName: \"kubernetes.io/projected/430249db-e9a3-4fdb-9090-96443b92a91b-kube-api-access-wg9f9\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.007482 master-0 kubenswrapper[10444]: E1205 10:37:50.007377 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit podName:430249db-e9a3-4fdb-9090-96443b92a91b nodeName:}" failed. No retries permitted until 2025-12-05 10:37:50.507356452 +0000 UTC m=+11.099269119 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit") pod "apiserver-7c8487d4d9-hsrsh" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b") : configmap "audit-0" not found Dec 05 10:37:50.007482 master-0 kubenswrapper[10444]: I1205 10:37:50.007446 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-config\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.007482 master-0 kubenswrapper[10444]: I1205 10:37:50.007474 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-client\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.007703 master-0 kubenswrapper[10444]: I1205 10:37:50.007665 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-encryption-config\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.007766 master-0 kubenswrapper[10444]: I1205 10:37:50.007724 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-audit-dir\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.007806 master-0 kubenswrapper[10444]: I1205 10:37:50.007786 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-audit-dir\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.008018 master-0 kubenswrapper[10444]: I1205 10:37:50.007940 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-node-pullsecrets\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.008145 master-0 kubenswrapper[10444]: I1205 10:37:50.008121 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-node-pullsecrets\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.008238 master-0 kubenswrapper[10444]: I1205 10:37:50.008201 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-image-import-ca\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.008316 master-0 kubenswrapper[10444]: I1205 10:37:50.008287 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-trusted-ca-bundle\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.008560 master-0 kubenswrapper[10444]: I1205 10:37:50.008517 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-config\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.008740 master-0 kubenswrapper[10444]: I1205 10:37:50.008705 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-serving-ca\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.022057 master-0 kubenswrapper[10444]: I1205 10:37:50.021991 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-encryption-config\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.022264 master-0 kubenswrapper[10444]: I1205 10:37:50.022145 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-client\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.025580 master-0 kubenswrapper[10444]: I1205 10:37:50.025521 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg9f9\" (UniqueName: \"kubernetes.io/projected/430249db-e9a3-4fdb-9090-96443b92a91b-kube-api-access-wg9f9\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.512590 master-0 kubenswrapper[10444]: I1205 10:37:50.512120 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.512764 master-0 kubenswrapper[10444]: E1205 10:37:50.512355 10444 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-0: configmap "audit-0" not found Dec 05 10:37:50.512764 master-0 kubenswrapper[10444]: I1205 10:37:50.512617 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:50.512764 master-0 kubenswrapper[10444]: E1205 10:37:50.512670 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit podName:430249db-e9a3-4fdb-9090-96443b92a91b nodeName:}" failed. No retries permitted until 2025-12-05 10:37:51.512647588 +0000 UTC m=+12.104560155 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit") pod "apiserver-7c8487d4d9-hsrsh" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b") : configmap "audit-0" not found Dec 05 10:37:50.512904 master-0 kubenswrapper[10444]: E1205 10:37:50.512854 10444 secret.go:189] Couldn't get secret openshift-apiserver/serving-cert: secret "serving-cert" not found Dec 05 10:37:50.513010 master-0 kubenswrapper[10444]: E1205 10:37:50.512978 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert podName:430249db-e9a3-4fdb-9090-96443b92a91b nodeName:}" failed. No retries permitted until 2025-12-05 10:37:51.512950857 +0000 UTC m=+12.104863454 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert") pod "apiserver-7c8487d4d9-hsrsh" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b") : secret "serving-cert" not found Dec 05 10:37:50.764764 master-0 kubenswrapper[10444]: I1205 10:37:50.764410 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" event={"ID":"82ef99d4-41b7-4859-a5a1-4e31ce614a2a","Type":"ContainerStarted","Data":"0f8b8ef2f443d099f39fe59fdd4477b3381b4efa500cd40773a08349e311ef2c"} Dec 05 10:37:50.767143 master-0 kubenswrapper[10444]: I1205 10:37:50.766925 10444 generic.go:334] "Generic (PLEG): container finished" podID="c22d947f-a5b6-4f24-b142-dd201c46293b" containerID="3e549526f5057891751a60c691572ec799806d7061304478b0c1e902a36e6208" exitCode=0 Dec 05 10:37:50.767582 master-0 kubenswrapper[10444]: I1205 10:37:50.767550 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" event={"ID":"c22d947f-a5b6-4f24-b142-dd201c46293b","Type":"ContainerDied","Data":"3e549526f5057891751a60c691572ec799806d7061304478b0c1e902a36e6208"} Dec 05 10:37:50.767854 master-0 kubenswrapper[10444]: I1205 10:37:50.767823 10444 scope.go:117] "RemoveContainer" containerID="3e549526f5057891751a60c691572ec799806d7061304478b0c1e902a36e6208" Dec 05 10:37:50.786224 master-0 kubenswrapper[10444]: I1205 10:37:50.785661 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:50.831437 master-0 kubenswrapper[10444]: I1205 10:37:50.831342 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h"] Dec 05 10:37:50.850294 master-0 kubenswrapper[10444]: I1205 10:37:50.850244 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb"] Dec 05 10:37:50.858102 master-0 kubenswrapper[10444]: W1205 10:37:50.858064 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7b29f89_e42d_4e53_ad14_05efdce933f0.slice/crio-a669846f3b9be1321aadc153fc634864e5c4e6b7c212feee982c065424fad087 WatchSource:0}: Error finding container a669846f3b9be1321aadc153fc634864e5c4e6b7c212feee982c065424fad087: Status 404 returned error can't find the container with id a669846f3b9be1321aadc153fc634864e5c4e6b7c212feee982c065424fad087 Dec 05 10:37:50.875128 master-0 kubenswrapper[10444]: I1205 10:37:50.875025 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-8649c48786-cgt5x"] Dec 05 10:37:50.936348 master-0 kubenswrapper[10444]: W1205 10:37:50.936316 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22676fac_b770_4937_9bee_7478bd1babb7.slice/crio-8066239534fa93e4ea7b428954f42f9cdca327545a8ae0d3c647f8ee2fd24e03 WatchSource:0}: Error finding container 8066239534fa93e4ea7b428954f42f9cdca327545a8ae0d3c647f8ee2fd24e03: Status 404 returned error can't find the container with id 8066239534fa93e4ea7b428954f42f9cdca327545a8ae0d3c647f8ee2fd24e03 Dec 05 10:37:51.097849 master-0 kubenswrapper[10444]: I1205 10:37:51.097229 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-node-tuning-operator/tuned-hvh88"] Dec 05 10:37:51.097849 master-0 kubenswrapper[10444]: I1205 10:37:51.097764 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.121568 master-0 kubenswrapper[10444]: I1205 10:37:51.121038 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:51.121568 master-0 kubenswrapper[10444]: E1205 10:37:51.121228 10444 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: configmap "client-ca" not found Dec 05 10:37:51.121568 master-0 kubenswrapper[10444]: E1205 10:37:51.121281 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca podName:61fd412b-c1c2-4902-9564-aab7a8b17845 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:59.121265185 +0000 UTC m=+19.713177772 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca") pod "controller-manager-6d9cb7b7fc-f9nz6" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845") : configmap "client-ca" not found Dec 05 10:37:51.222468 master-0 kubenswrapper[10444]: I1205 10:37:51.222370 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-lib-modules\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.222655 master-0 kubenswrapper[10444]: I1205 10:37:51.222553 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-modprobe-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.222711 master-0 kubenswrapper[10444]: I1205 10:37:51.222654 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.222985 master-0 kubenswrapper[10444]: I1205 10:37:51.222955 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsrkp\" (UniqueName: \"kubernetes.io/projected/33616deb-ea10-4a38-8681-ab023b526b11-kube-api-access-rsrkp\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.223035 master-0 kubenswrapper[10444]: I1205 10:37:51.222998 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-run\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.223067 master-0 kubenswrapper[10444]: I1205 10:37:51.223045 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-conf\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.223067 master-0 kubenswrapper[10444]: I1205 10:37:51.223063 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-var-lib-kubelet\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.223122 master-0 kubenswrapper[10444]: I1205 10:37:51.223110 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysconfig\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.223161 master-0 kubenswrapper[10444]: I1205 10:37:51.223146 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-tmp\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.223193 master-0 kubenswrapper[10444]: I1205 10:37:51.223165 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-systemd\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.223193 master-0 kubenswrapper[10444]: I1205 10:37:51.223188 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-kubernetes\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.223410 master-0 kubenswrapper[10444]: I1205 10:37:51.223370 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-sys\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.223624 master-0 kubenswrapper[10444]: I1205 10:37:51.223572 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-host\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.223674 master-0 kubenswrapper[10444]: I1205 10:37:51.223656 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-etc-tuned\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.324802 master-0 kubenswrapper[10444]: I1205 10:37:51.324753 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-conf\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.324802 master-0 kubenswrapper[10444]: I1205 10:37:51.324793 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-var-lib-kubelet\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325015 master-0 kubenswrapper[10444]: I1205 10:37:51.324834 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysconfig\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325062 master-0 kubenswrapper[10444]: I1205 10:37:51.325038 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-conf\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325125 master-0 kubenswrapper[10444]: I1205 10:37:51.325102 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-tmp\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325163 master-0 kubenswrapper[10444]: I1205 10:37:51.325141 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-systemd\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325192 master-0 kubenswrapper[10444]: I1205 10:37:51.325164 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-kubernetes\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325192 master-0 kubenswrapper[10444]: I1205 10:37:51.325185 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-sys\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325305 master-0 kubenswrapper[10444]: I1205 10:37:51.325211 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-host\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325305 master-0 kubenswrapper[10444]: I1205 10:37:51.325236 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-etc-tuned\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325305 master-0 kubenswrapper[10444]: I1205 10:37:51.325297 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-lib-modules\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325401 master-0 kubenswrapper[10444]: I1205 10:37:51.325336 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-modprobe-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325401 master-0 kubenswrapper[10444]: I1205 10:37:51.325372 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325488 master-0 kubenswrapper[10444]: I1205 10:37:51.325453 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsrkp\" (UniqueName: \"kubernetes.io/projected/33616deb-ea10-4a38-8681-ab023b526b11-kube-api-access-rsrkp\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325488 master-0 kubenswrapper[10444]: I1205 10:37:51.325478 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-run\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325712 master-0 kubenswrapper[10444]: I1205 10:37:51.325675 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysconfig\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325773 master-0 kubenswrapper[10444]: I1205 10:37:51.325728 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-modprobe-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325812 master-0 kubenswrapper[10444]: I1205 10:37:51.325744 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-kubernetes\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325812 master-0 kubenswrapper[10444]: I1205 10:37:51.325784 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-sys\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325901 master-0 kubenswrapper[10444]: I1205 10:37:51.325867 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-var-lib-kubelet\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.325944 master-0 kubenswrapper[10444]: I1205 10:37:51.325919 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-systemd\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.326054 master-0 kubenswrapper[10444]: I1205 10:37:51.326024 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-host\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.326243 master-0 kubenswrapper[10444]: I1205 10:37:51.326206 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.326243 master-0 kubenswrapper[10444]: I1205 10:37:51.326227 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-lib-modules\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.326571 master-0 kubenswrapper[10444]: I1205 10:37:51.326210 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-run\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.337714 master-0 kubenswrapper[10444]: I1205 10:37:51.337673 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-tmp\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.337821 master-0 kubenswrapper[10444]: I1205 10:37:51.337737 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-etc-tuned\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.340777 master-0 kubenswrapper[10444]: I1205 10:37:51.340735 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsrkp\" (UniqueName: \"kubernetes.io/projected/33616deb-ea10-4a38-8681-ab023b526b11-kube-api-access-rsrkp\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.417940 master-0 kubenswrapper[10444]: I1205 10:37:51.417833 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:37:51.442650 master-0 kubenswrapper[10444]: W1205 10:37:51.442616 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod33616deb_ea10_4a38_8681_ab023b526b11.slice/crio-c7591d56b2262374531f713ccc5ed133dec6c53f18edc5e2a849a4fb0c7ac26d WatchSource:0}: Error finding container c7591d56b2262374531f713ccc5ed133dec6c53f18edc5e2a849a4fb0c7ac26d: Status 404 returned error can't find the container with id c7591d56b2262374531f713ccc5ed133dec6c53f18edc5e2a849a4fb0c7ac26d Dec 05 10:37:51.473845 master-0 kubenswrapper[10444]: I1205 10:37:51.473718 10444 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:51.528416 master-0 kubenswrapper[10444]: I1205 10:37:51.528235 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:51.528680 master-0 kubenswrapper[10444]: E1205 10:37:51.528467 10444 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-0: configmap "audit-0" not found Dec 05 10:37:51.528680 master-0 kubenswrapper[10444]: E1205 10:37:51.528567 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit podName:430249db-e9a3-4fdb-9090-96443b92a91b nodeName:}" failed. No retries permitted until 2025-12-05 10:37:53.528540379 +0000 UTC m=+14.120453126 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit") pod "apiserver-7c8487d4d9-hsrsh" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b") : configmap "audit-0" not found Dec 05 10:37:51.528680 master-0 kubenswrapper[10444]: I1205 10:37:51.528658 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:51.528797 master-0 kubenswrapper[10444]: E1205 10:37:51.528782 10444 secret.go:189] Couldn't get secret openshift-apiserver/serving-cert: secret "serving-cert" not found Dec 05 10:37:51.528860 master-0 kubenswrapper[10444]: E1205 10:37:51.528841 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert podName:430249db-e9a3-4fdb-9090-96443b92a91b nodeName:}" failed. No retries permitted until 2025-12-05 10:37:53.528828167 +0000 UTC m=+14.120740914 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert") pod "apiserver-7c8487d4d9-hsrsh" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b") : secret "serving-cert" not found Dec 05 10:37:51.628870 master-0 kubenswrapper[10444]: I1205 10:37:51.628783 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:37:51.778082 master-0 kubenswrapper[10444]: I1205 10:37:51.777983 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" event={"ID":"c22d947f-a5b6-4f24-b142-dd201c46293b","Type":"ContainerStarted","Data":"fbae6ca1248953a89b3c9c60d625d45cf8e23affac69834bf8dc12ad818c8d4c"} Dec 05 10:37:51.778513 master-0 kubenswrapper[10444]: I1205 10:37:51.778493 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:51.779686 master-0 kubenswrapper[10444]: I1205 10:37:51.779650 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" event={"ID":"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f","Type":"ContainerStarted","Data":"4308a58237214eca1f5ba917e53d41b87b4e41bf19760a2305217ef9ccb8ed3f"} Dec 05 10:37:51.783767 master-0 kubenswrapper[10444]: I1205 10:37:51.783714 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" event={"ID":"22676fac-b770-4937-9bee-7478bd1babb7","Type":"ContainerStarted","Data":"8066239534fa93e4ea7b428954f42f9cdca327545a8ae0d3c647f8ee2fd24e03"} Dec 05 10:37:51.785820 master-0 kubenswrapper[10444]: I1205 10:37:51.785791 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" event={"ID":"eb290494-a456-4f0e-9afc-f20abab1a1bf","Type":"ContainerStarted","Data":"ed7f4ad9d55acd590e0a6b8f43c87e139f3a69e3b2e25fa90f8eb572ae8d52e6"} Dec 05 10:37:51.787625 master-0 kubenswrapper[10444]: I1205 10:37:51.787590 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-hvh88" event={"ID":"33616deb-ea10-4a38-8681-ab023b526b11","Type":"ContainerStarted","Data":"20698f31b60e6e2c7bf3040ce4e7430ceb9edd49bb5ccef0787cf77ed8fa89d4"} Dec 05 10:37:51.787625 master-0 kubenswrapper[10444]: I1205 10:37:51.787616 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-node-tuning-operator/tuned-hvh88" event={"ID":"33616deb-ea10-4a38-8681-ab023b526b11","Type":"ContainerStarted","Data":"c7591d56b2262374531f713ccc5ed133dec6c53f18edc5e2a849a4fb0c7ac26d"} Dec 05 10:37:51.788832 master-0 kubenswrapper[10444]: I1205 10:37:51.788798 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" event={"ID":"5a1bdc70-6412-47e0-8330-04d796cc8d55","Type":"ContainerStarted","Data":"0d874913b26ce96decb20e6d175c667c355d6a65334e625b074f3811b11e3027"} Dec 05 10:37:51.790273 master-0 kubenswrapper[10444]: I1205 10:37:51.790245 10444 generic.go:334] "Generic (PLEG): container finished" podID="49051e6e-5a2f-45c8-bad0-374514a91c07" containerID="8a2b637ea3b720bb56a1f54c7baf3eb708da58746d9f80b01f4fc142a426b6b4" exitCode=0 Dec 05 10:37:51.790344 master-0 kubenswrapper[10444]: I1205 10:37:51.790284 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" event={"ID":"49051e6e-5a2f-45c8-bad0-374514a91c07","Type":"ContainerDied","Data":"8a2b637ea3b720bb56a1f54c7baf3eb708da58746d9f80b01f4fc142a426b6b4"} Dec 05 10:37:51.791761 master-0 kubenswrapper[10444]: I1205 10:37:51.791658 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" event={"ID":"f7b29f89-e42d-4e53-ad14-05efdce933f0","Type":"ContainerStarted","Data":"a669846f3b9be1321aadc153fc634864e5c4e6b7c212feee982c065424fad087"} Dec 05 10:37:51.858054 master-0 kubenswrapper[10444]: I1205 10:37:51.857952 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-node-tuning-operator/tuned-hvh88" podStartSLOduration=0.857853649 podStartE2EDuration="857.853649ms" podCreationTimestamp="2025-12-05 10:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:37:51.854059581 +0000 UTC m=+12.445972168" watchObservedRunningTime="2025-12-05 10:37:51.857853649 +0000 UTC m=+12.449766216" Dec 05 10:37:53.568438 master-0 kubenswrapper[10444]: I1205 10:37:53.567978 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:53.569135 master-0 kubenswrapper[10444]: I1205 10:37:53.568494 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit\") pod \"apiserver-7c8487d4d9-hsrsh\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:53.569135 master-0 kubenswrapper[10444]: E1205 10:37:53.568614 10444 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-0: configmap "audit-0" not found Dec 05 10:37:53.569135 master-0 kubenswrapper[10444]: E1205 10:37:53.568674 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit podName:430249db-e9a3-4fdb-9090-96443b92a91b nodeName:}" failed. No retries permitted until 2025-12-05 10:37:57.568653867 +0000 UTC m=+18.160566434 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit") pod "apiserver-7c8487d4d9-hsrsh" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b") : configmap "audit-0" not found Dec 05 10:37:53.569702 master-0 kubenswrapper[10444]: E1205 10:37:53.569662 10444 secret.go:189] Couldn't get secret openshift-apiserver/serving-cert: secret "serving-cert" not found Dec 05 10:37:53.569755 master-0 kubenswrapper[10444]: E1205 10:37:53.569748 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert podName:430249db-e9a3-4fdb-9090-96443b92a91b nodeName:}" failed. No retries permitted until 2025-12-05 10:37:57.569729288 +0000 UTC m=+18.161641855 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert") pod "apiserver-7c8487d4d9-hsrsh" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b") : secret "serving-cert" not found Dec 05 10:37:53.650787 master-0 kubenswrapper[10444]: I1205 10:37:53.649625 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-apiserver/apiserver-7c8487d4d9-hsrsh"] Dec 05 10:37:53.650787 master-0 kubenswrapper[10444]: E1205 10:37:53.649866 10444 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[audit serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" podUID="430249db-e9a3-4fdb-9090-96443b92a91b" Dec 05 10:37:53.792538 master-0 kubenswrapper[10444]: I1205 10:37:53.792451 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:37:53.806164 master-0 kubenswrapper[10444]: I1205 10:37:53.806114 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:53.813640 master-0 kubenswrapper[10444]: I1205 10:37:53.813612 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:53.974935 master-0 kubenswrapper[10444]: I1205 10:37:53.974465 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-config\") pod \"430249db-e9a3-4fdb-9090-96443b92a91b\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " Dec 05 10:37:53.974935 master-0 kubenswrapper[10444]: I1205 10:37:53.974518 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-trusted-ca-bundle\") pod \"430249db-e9a3-4fdb-9090-96443b92a91b\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " Dec 05 10:37:53.974935 master-0 kubenswrapper[10444]: I1205 10:37:53.974545 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-client\") pod \"430249db-e9a3-4fdb-9090-96443b92a91b\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " Dec 05 10:37:53.974935 master-0 kubenswrapper[10444]: I1205 10:37:53.974565 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg9f9\" (UniqueName: \"kubernetes.io/projected/430249db-e9a3-4fdb-9090-96443b92a91b-kube-api-access-wg9f9\") pod \"430249db-e9a3-4fdb-9090-96443b92a91b\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " Dec 05 10:37:53.974935 master-0 kubenswrapper[10444]: I1205 10:37:53.974588 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-audit-dir\") pod \"430249db-e9a3-4fdb-9090-96443b92a91b\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " Dec 05 10:37:53.974935 master-0 kubenswrapper[10444]: I1205 10:37:53.974606 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-image-import-ca\") pod \"430249db-e9a3-4fdb-9090-96443b92a91b\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " Dec 05 10:37:53.974935 master-0 kubenswrapper[10444]: I1205 10:37:53.974630 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-encryption-config\") pod \"430249db-e9a3-4fdb-9090-96443b92a91b\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " Dec 05 10:37:53.974935 master-0 kubenswrapper[10444]: I1205 10:37:53.974671 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-serving-ca\") pod \"430249db-e9a3-4fdb-9090-96443b92a91b\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " Dec 05 10:37:53.974935 master-0 kubenswrapper[10444]: I1205 10:37:53.974698 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-node-pullsecrets\") pod \"430249db-e9a3-4fdb-9090-96443b92a91b\" (UID: \"430249db-e9a3-4fdb-9090-96443b92a91b\") " Dec 05 10:37:53.974935 master-0 kubenswrapper[10444]: I1205 10:37:53.974690 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "430249db-e9a3-4fdb-9090-96443b92a91b" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:53.974935 master-0 kubenswrapper[10444]: I1205 10:37:53.974942 10444 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-audit-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:53.975493 master-0 kubenswrapper[10444]: I1205 10:37:53.974993 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "430249db-e9a3-4fdb-9090-96443b92a91b" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:37:53.975493 master-0 kubenswrapper[10444]: I1205 10:37:53.975174 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "430249db-e9a3-4fdb-9090-96443b92a91b" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:37:53.975493 master-0 kubenswrapper[10444]: I1205 10:37:53.975226 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "430249db-e9a3-4fdb-9090-96443b92a91b" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:37:53.975493 master-0 kubenswrapper[10444]: I1205 10:37:53.975334 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "430249db-e9a3-4fdb-9090-96443b92a91b" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:37:53.975854 master-0 kubenswrapper[10444]: I1205 10:37:53.975814 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-config" (OuterVolumeSpecName: "config") pod "430249db-e9a3-4fdb-9090-96443b92a91b" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:37:53.982304 master-0 kubenswrapper[10444]: I1205 10:37:53.979321 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/430249db-e9a3-4fdb-9090-96443b92a91b-kube-api-access-wg9f9" (OuterVolumeSpecName: "kube-api-access-wg9f9") pod "430249db-e9a3-4fdb-9090-96443b92a91b" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b"). InnerVolumeSpecName "kube-api-access-wg9f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:37:53.982304 master-0 kubenswrapper[10444]: I1205 10:37:53.980023 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "430249db-e9a3-4fdb-9090-96443b92a91b" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:37:53.983529 master-0 kubenswrapper[10444]: I1205 10:37:53.983481 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "430249db-e9a3-4fdb-9090-96443b92a91b" (UID: "430249db-e9a3-4fdb-9090-96443b92a91b"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:37:54.076123 master-0 kubenswrapper[10444]: I1205 10:37:54.076025 10444 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-image-import-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:54.076123 master-0 kubenswrapper[10444]: I1205 10:37:54.076108 10444 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-encryption-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:54.076123 master-0 kubenswrapper[10444]: I1205 10:37:54.076121 10444 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-serving-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:54.076123 master-0 kubenswrapper[10444]: I1205 10:37:54.076134 10444 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/430249db-e9a3-4fdb-9090-96443b92a91b-node-pullsecrets\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:54.076509 master-0 kubenswrapper[10444]: I1205 10:37:54.076147 10444 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:54.076509 master-0 kubenswrapper[10444]: I1205 10:37:54.076160 10444 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-trusted-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:54.076509 master-0 kubenswrapper[10444]: I1205 10:37:54.076173 10444 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-etcd-client\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:54.076509 master-0 kubenswrapper[10444]: I1205 10:37:54.076186 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg9f9\" (UniqueName: \"kubernetes.io/projected/430249db-e9a3-4fdb-9090-96443b92a91b-kube-api-access-wg9f9\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:54.301393 master-0 kubenswrapper[10444]: I1205 10:37:54.301254 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/installer-1-master-0"] Dec 05 10:37:54.302532 master-0 kubenswrapper[10444]: I1205 10:37:54.302033 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:37:54.304186 master-0 kubenswrapper[10444]: I1205 10:37:54.304064 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler"/"kube-root-ca.crt" Dec 05 10:37:54.313642 master-0 kubenswrapper[10444]: I1205 10:37:54.313407 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/installer-1-master-0"] Dec 05 10:37:54.334245 master-0 kubenswrapper[10444]: I1205 10:37:54.334151 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kube-api-access\") pod \"installer-1-master-0\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:37:54.334414 master-0 kubenswrapper[10444]: I1205 10:37:54.334362 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-var-lock\") pod \"installer-1-master-0\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:37:54.334521 master-0 kubenswrapper[10444]: I1205 10:37:54.334473 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:37:54.444862 master-0 kubenswrapper[10444]: I1205 10:37:54.435827 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kube-api-access\") pod \"installer-1-master-0\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:37:54.444862 master-0 kubenswrapper[10444]: I1205 10:37:54.435919 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-var-lock\") pod \"installer-1-master-0\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:37:54.444862 master-0 kubenswrapper[10444]: I1205 10:37:54.435983 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:37:54.444862 master-0 kubenswrapper[10444]: I1205 10:37:54.436064 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:37:54.444862 master-0 kubenswrapper[10444]: I1205 10:37:54.436362 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-var-lock\") pod \"installer-1-master-0\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:37:54.459328 master-0 kubenswrapper[10444]: I1205 10:37:54.459268 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kube-api-access\") pod \"installer-1-master-0\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:37:54.620803 master-0 kubenswrapper[10444]: I1205 10:37:54.620681 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:37:54.808586 master-0 kubenswrapper[10444]: I1205 10:37:54.808533 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-7c8487d4d9-hsrsh" Dec 05 10:37:54.897545 master-0 kubenswrapper[10444]: I1205 10:37:54.896727 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-5b9fd577f8-6sxcx"] Dec 05 10:37:54.897545 master-0 kubenswrapper[10444]: I1205 10:37:54.897520 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.902365 master-0 kubenswrapper[10444]: I1205 10:37:54.901999 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 10:37:54.902365 master-0 kubenswrapper[10444]: I1205 10:37:54.902074 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 10:37:54.902365 master-0 kubenswrapper[10444]: I1205 10:37:54.902126 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 10:37:54.902365 master-0 kubenswrapper[10444]: I1205 10:37:54.902154 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 10:37:54.902365 master-0 kubenswrapper[10444]: I1205 10:37:54.902187 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 10:37:54.902365 master-0 kubenswrapper[10444]: I1205 10:37:54.902225 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 10:37:54.902365 master-0 kubenswrapper[10444]: I1205 10:37:54.902196 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 10:37:54.902365 master-0 kubenswrapper[10444]: I1205 10:37:54.902251 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 10:37:54.902365 master-0 kubenswrapper[10444]: I1205 10:37:54.902272 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 10:37:54.905491 master-0 kubenswrapper[10444]: I1205 10:37:54.905409 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-apiserver/apiserver-7c8487d4d9-hsrsh"] Dec 05 10:37:54.913580 master-0 kubenswrapper[10444]: I1205 10:37:54.913454 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 10:37:54.913755 master-0 kubenswrapper[10444]: I1205 10:37:54.913589 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-5b9fd577f8-6sxcx"] Dec 05 10:37:54.914245 master-0 kubenswrapper[10444]: I1205 10:37:54.914201 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-apiserver/apiserver-7c8487d4d9-hsrsh"] Dec 05 10:37:54.945275 master-0 kubenswrapper[10444]: I1205 10:37:54.945057 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-serving-ca\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.945275 master-0 kubenswrapper[10444]: I1205 10:37:54.945125 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.945275 master-0 kubenswrapper[10444]: I1205 10:37:54.945183 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-config\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.945275 master-0 kubenswrapper[10444]: I1205 10:37:54.945206 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.945275 master-0 kubenswrapper[10444]: I1205 10:37:54.945241 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-client\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.945655 master-0 kubenswrapper[10444]: I1205 10:37:54.945415 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-image-import-ca\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.945655 master-0 kubenswrapper[10444]: I1205 10:37:54.945537 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-encryption-config\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.945655 master-0 kubenswrapper[10444]: I1205 10:37:54.945583 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m88gw\" (UniqueName: \"kubernetes.io/projected/d5198438-06ae-4e63-a7e3-950ba23bba9c-kube-api-access-m88gw\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.945759 master-0 kubenswrapper[10444]: I1205 10:37:54.945667 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit-dir\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.945759 master-0 kubenswrapper[10444]: I1205 10:37:54.945717 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-trusted-ca-bundle\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.945842 master-0 kubenswrapper[10444]: I1205 10:37:54.945778 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-node-pullsecrets\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:54.945842 master-0 kubenswrapper[10444]: I1205 10:37:54.945838 10444 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/430249db-e9a3-4fdb-9090-96443b92a91b-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:54.945915 master-0 kubenswrapper[10444]: I1205 10:37:54.945859 10444 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/430249db-e9a3-4fdb-9090-96443b92a91b-audit\") on node \"master-0\" DevicePath \"\"" Dec 05 10:37:55.046998 master-0 kubenswrapper[10444]: I1205 10:37:55.046907 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-config\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.046998 master-0 kubenswrapper[10444]: I1205 10:37:55.046968 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.046998 master-0 kubenswrapper[10444]: I1205 10:37:55.047009 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-client\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.047349 master-0 kubenswrapper[10444]: I1205 10:37:55.047059 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-image-import-ca\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.047349 master-0 kubenswrapper[10444]: I1205 10:37:55.047092 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-encryption-config\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.047349 master-0 kubenswrapper[10444]: I1205 10:37:55.047121 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m88gw\" (UniqueName: \"kubernetes.io/projected/d5198438-06ae-4e63-a7e3-950ba23bba9c-kube-api-access-m88gw\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.047349 master-0 kubenswrapper[10444]: I1205 10:37:55.047169 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit-dir\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.047349 master-0 kubenswrapper[10444]: I1205 10:37:55.047194 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-trusted-ca-bundle\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.047349 master-0 kubenswrapper[10444]: I1205 10:37:55.047228 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-node-pullsecrets\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.047349 master-0 kubenswrapper[10444]: I1205 10:37:55.047257 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-serving-ca\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.047349 master-0 kubenswrapper[10444]: I1205 10:37:55.047301 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.050403 master-0 kubenswrapper[10444]: E1205 10:37:55.048013 10444 secret.go:189] Couldn't get secret openshift-apiserver/serving-cert: secret "serving-cert" not found Dec 05 10:37:55.050403 master-0 kubenswrapper[10444]: E1205 10:37:55.048101 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert podName:d5198438-06ae-4e63-a7e3-950ba23bba9c nodeName:}" failed. No retries permitted until 2025-12-05 10:37:55.548077204 +0000 UTC m=+16.139989851 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert") pod "apiserver-5b9fd577f8-6sxcx" (UID: "d5198438-06ae-4e63-a7e3-950ba23bba9c") : secret "serving-cert" not found Dec 05 10:37:55.050403 master-0 kubenswrapper[10444]: I1205 10:37:55.048107 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.050403 master-0 kubenswrapper[10444]: I1205 10:37:55.048566 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-image-import-ca\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.050403 master-0 kubenswrapper[10444]: I1205 10:37:55.049122 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-trusted-ca-bundle\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.050403 master-0 kubenswrapper[10444]: I1205 10:37:55.049157 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit-dir\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.050403 master-0 kubenswrapper[10444]: I1205 10:37:55.049195 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-node-pullsecrets\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.051052 master-0 kubenswrapper[10444]: I1205 10:37:55.050968 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-config\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.051052 master-0 kubenswrapper[10444]: I1205 10:37:55.051046 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-serving-ca\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.053664 master-0 kubenswrapper[10444]: I1205 10:37:55.053594 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-encryption-config\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.053957 master-0 kubenswrapper[10444]: I1205 10:37:55.053914 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-client\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.068181 master-0 kubenswrapper[10444]: I1205 10:37:55.068062 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m88gw\" (UniqueName: \"kubernetes.io/projected/d5198438-06ae-4e63-a7e3-950ba23bba9c-kube-api-access-m88gw\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.553903 master-0 kubenswrapper[10444]: I1205 10:37:55.553860 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:55.554389 master-0 kubenswrapper[10444]: E1205 10:37:55.554365 10444 secret.go:189] Couldn't get secret openshift-apiserver/serving-cert: secret "serving-cert" not found Dec 05 10:37:55.554534 master-0 kubenswrapper[10444]: E1205 10:37:55.554520 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert podName:d5198438-06ae-4e63-a7e3-950ba23bba9c nodeName:}" failed. No retries permitted until 2025-12-05 10:37:56.55449977 +0000 UTC m=+17.146412337 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert") pod "apiserver-5b9fd577f8-6sxcx" (UID: "d5198438-06ae-4e63-a7e3-950ba23bba9c") : secret "serving-cert" not found Dec 05 10:37:55.907128 master-0 kubenswrapper[10444]: I1205 10:37:55.906998 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="430249db-e9a3-4fdb-9090-96443b92a91b" path="/var/lib/kubelet/pods/430249db-e9a3-4fdb-9090-96443b92a91b/volumes" Dec 05 10:37:56.313872 master-0 kubenswrapper[10444]: I1205 10:37:56.312449 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:56.313872 master-0 kubenswrapper[10444]: I1205 10:37:56.312541 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:56.313872 master-0 kubenswrapper[10444]: I1205 10:37:56.312582 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:56.313872 master-0 kubenswrapper[10444]: I1205 10:37:56.312608 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:56.313872 master-0 kubenswrapper[10444]: I1205 10:37:56.312679 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:56.313872 master-0 kubenswrapper[10444]: I1205 10:37:56.312700 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:56.318882 master-0 kubenswrapper[10444]: I1205 10:37:56.318712 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:56.319486 master-0 kubenswrapper[10444]: I1205 10:37:56.319407 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:56.319877 master-0 kubenswrapper[10444]: I1205 10:37:56.319818 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:56.320056 master-0 kubenswrapper[10444]: I1205 10:37:56.320014 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:56.320134 master-0 kubenswrapper[10444]: I1205 10:37:56.320091 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:56.321321 master-0 kubenswrapper[10444]: I1205 10:37:56.321280 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"multus-admission-controller-7dfc5b745f-67rx7\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:56.413312 master-0 kubenswrapper[10444]: I1205 10:37:56.413192 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:56.413312 master-0 kubenswrapper[10444]: I1205 10:37:56.413262 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:56.413617 master-0 kubenswrapper[10444]: E1205 10:37:56.413515 10444 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: configmap "client-ca" not found Dec 05 10:37:56.413617 master-0 kubenswrapper[10444]: E1205 10:37:56.413607 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca podName:98fd575c-ffc4-44c2-9947-c18d3f53e2c9 nodeName:}" failed. No retries permitted until 2025-12-05 10:38:12.413579909 +0000 UTC m=+33.005492476 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca") pod "route-controller-manager-76d4564964-xm2tr" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9") : configmap "client-ca" not found Dec 05 10:37:56.418387 master-0 kubenswrapper[10444]: I1205 10:37:56.418339 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert\") pod \"route-controller-manager-76d4564964-xm2tr\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:37:56.440715 master-0 kubenswrapper[10444]: I1205 10:37:56.440635 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:37:56.440917 master-0 kubenswrapper[10444]: I1205 10:37:56.440658 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:37:56.440917 master-0 kubenswrapper[10444]: I1205 10:37:56.440762 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:37:56.440917 master-0 kubenswrapper[10444]: I1205 10:37:56.440861 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:37:56.466228 master-0 kubenswrapper[10444]: I1205 10:37:56.466166 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:37:56.466890 master-0 kubenswrapper[10444]: I1205 10:37:56.466831 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:37:56.615395 master-0 kubenswrapper[10444]: I1205 10:37:56.615269 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:56.615620 master-0 kubenswrapper[10444]: E1205 10:37:56.615472 10444 secret.go:189] Couldn't get secret openshift-apiserver/serving-cert: secret "serving-cert" not found Dec 05 10:37:56.615620 master-0 kubenswrapper[10444]: E1205 10:37:56.615523 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert podName:d5198438-06ae-4e63-a7e3-950ba23bba9c nodeName:}" failed. No retries permitted until 2025-12-05 10:37:58.615507157 +0000 UTC m=+19.207419724 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert") pod "apiserver-5b9fd577f8-6sxcx" (UID: "d5198438-06ae-4e63-a7e3-950ba23bba9c") : secret "serving-cert" not found Dec 05 10:37:56.941674 master-0 kubenswrapper[10444]: I1205 10:37:56.941583 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" event={"ID":"82ef99d4-41b7-4859-a5a1-4e31ce614a2a","Type":"ContainerStarted","Data":"8575f477d29594718db712e5d56ee7bf4ba0a8c565ae4540b71d0eb7f5d45356"} Dec 05 10:37:57.015615 master-0 kubenswrapper[10444]: I1205 10:37:57.010220 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/installer-1-master-0"] Dec 05 10:37:57.165006 master-0 kubenswrapper[10444]: I1205 10:37:57.164900 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8gjgm"] Dec 05 10:37:57.170847 master-0 kubenswrapper[10444]: I1205 10:37:57.170802 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk"] Dec 05 10:37:57.175215 master-0 kubenswrapper[10444]: W1205 10:37:57.175158 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a1e2bf3_2e33_4a2a_b306_6d5e1d29727c.slice/crio-62e34bf231128449a630a3dad98495dd55715547e65586a815afe200566d05a0 WatchSource:0}: Error finding container 62e34bf231128449a630a3dad98495dd55715547e65586a815afe200566d05a0: Status 404 returned error can't find the container with id 62e34bf231128449a630a3dad98495dd55715547e65586a815afe200566d05a0 Dec 05 10:37:57.179408 master-0 kubenswrapper[10444]: W1205 10:37:57.179289 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46c5ace3_2d9e_40b3_a9ab_fd0ff8b0b082.slice/crio-0e02576ddbd856f2a1c30d4bbdcb0b27384c73dfe012be69bbb5c2168f8df7fd WatchSource:0}: Error finding container 0e02576ddbd856f2a1c30d4bbdcb0b27384c73dfe012be69bbb5c2168f8df7fd: Status 404 returned error can't find the container with id 0e02576ddbd856f2a1c30d4bbdcb0b27384c73dfe012be69bbb5c2168f8df7fd Dec 05 10:37:57.194158 master-0 kubenswrapper[10444]: I1205 10:37:57.193937 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn"] Dec 05 10:37:57.207797 master-0 kubenswrapper[10444]: W1205 10:37:57.207755 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c649a16_c187_412e_b5da_62a00bee38ab.slice/crio-77b83e99ece1e07058ac092ebb18d76d85bb5a90292440cd154e967f63253776 WatchSource:0}: Error finding container 77b83e99ece1e07058ac092ebb18d76d85bb5a90292440cd154e967f63253776: Status 404 returned error can't find the container with id 77b83e99ece1e07058ac092ebb18d76d85bb5a90292440cd154e967f63253776 Dec 05 10:37:57.304358 master-0 kubenswrapper[10444]: I1205 10:37:57.301040 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-f797b99b6-z9qcl"] Dec 05 10:37:57.312256 master-0 kubenswrapper[10444]: I1205 10:37:57.312203 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr"] Dec 05 10:37:57.317510 master-0 kubenswrapper[10444]: W1205 10:37:57.317458 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod926263c4_ec5b_41cb_9c30_0c88f636035f.slice/crio-ba6a43bb61d242c8e032f7cab8a9cf0e806cdf03c0a9ce44ec6c444a2216f39a WatchSource:0}: Error finding container ba6a43bb61d242c8e032f7cab8a9cf0e806cdf03c0a9ce44ec6c444a2216f39a: Status 404 returned error can't find the container with id ba6a43bb61d242c8e032f7cab8a9cf0e806cdf03c0a9ce44ec6c444a2216f39a Dec 05 10:37:57.322968 master-0 kubenswrapper[10444]: I1205 10:37:57.322357 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-7dfc5b745f-67rx7"] Dec 05 10:37:57.336731 master-0 kubenswrapper[10444]: W1205 10:37:57.327918 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda722cda9_29a0_4b7f_8e1d_9a8950ed765a.slice/crio-aa871dfa794fea5d166618a43475fe935ef7d4d3799afbc3d92debffe4b76081 WatchSource:0}: Error finding container aa871dfa794fea5d166618a43475fe935ef7d4d3799afbc3d92debffe4b76081: Status 404 returned error can't find the container with id aa871dfa794fea5d166618a43475fe935ef7d4d3799afbc3d92debffe4b76081 Dec 05 10:37:57.347146 master-0 kubenswrapper[10444]: W1205 10:37:57.338041 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d76404b_6d62_4a61_b6f6_0c8073eba198.slice/crio-e9475cd437d6e0adb87102a00093bc37d7716f625ab1f62c89982010f22e3b2c WatchSource:0}: Error finding container e9475cd437d6e0adb87102a00093bc37d7716f625ab1f62c89982010f22e3b2c: Status 404 returned error can't find the container with id e9475cd437d6e0adb87102a00093bc37d7716f625ab1f62c89982010f22e3b2c Dec 05 10:37:57.659736 master-0 kubenswrapper[10444]: I1205 10:37:57.659678 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-4vxng"] Dec 05 10:37:57.660547 master-0 kubenswrapper[10444]: I1205 10:37:57.660508 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:57.669045 master-0 kubenswrapper[10444]: I1205 10:37:57.668976 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 10:37:57.669287 master-0 kubenswrapper[10444]: I1205 10:37:57.669264 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 10:37:57.669444 master-0 kubenswrapper[10444]: I1205 10:37:57.669401 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 10:37:57.669621 master-0 kubenswrapper[10444]: I1205 10:37:57.669585 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 10:37:57.677527 master-0 kubenswrapper[10444]: I1205 10:37:57.675936 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4vxng"] Dec 05 10:37:57.828947 master-0 kubenswrapper[10444]: I1205 10:37:57.827903 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/installer-1-master-0"] Dec 05 10:37:57.828947 master-0 kubenswrapper[10444]: I1205 10:37:57.828408 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-1-master-0" Dec 05 10:37:57.829316 master-0 kubenswrapper[10444]: I1205 10:37:57.829221 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:57.829389 master-0 kubenswrapper[10444]: I1205 10:37:57.829315 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx7jr\" (UniqueName: \"kubernetes.io/projected/58baad85-de54-49e7-a13e-d470d9c50d11-kube-api-access-jx7jr\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:57.829389 master-0 kubenswrapper[10444]: I1205 10:37:57.829345 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58baad85-de54-49e7-a13e-d470d9c50d11-config-volume\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:57.833200 master-0 kubenswrapper[10444]: I1205 10:37:57.833054 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd"/"kube-root-ca.crt" Dec 05 10:37:57.833946 master-0 kubenswrapper[10444]: I1205 10:37:57.833861 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd/installer-1-master-0"] Dec 05 10:37:57.930298 master-0 kubenswrapper[10444]: I1205 10:37:57.930231 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx7jr\" (UniqueName: \"kubernetes.io/projected/58baad85-de54-49e7-a13e-d470d9c50d11-kube-api-access-jx7jr\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:57.930298 master-0 kubenswrapper[10444]: I1205 10:37:57.930290 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58baad85-de54-49e7-a13e-d470d9c50d11-config-volume\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:57.931381 master-0 kubenswrapper[10444]: I1205 10:37:57.930868 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kube-api-access\") pod \"installer-1-master-0\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " pod="openshift-etcd/installer-1-master-0" Dec 05 10:37:57.932171 master-0 kubenswrapper[10444]: I1205 10:37:57.932125 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58baad85-de54-49e7-a13e-d470d9c50d11-config-volume\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:57.932413 master-0 kubenswrapper[10444]: I1205 10:37:57.932193 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:57.932413 master-0 kubenswrapper[10444]: I1205 10:37:57.932222 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " pod="openshift-etcd/installer-1-master-0" Dec 05 10:37:57.932413 master-0 kubenswrapper[10444]: I1205 10:37:57.932319 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-var-lock\") pod \"installer-1-master-0\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " pod="openshift-etcd/installer-1-master-0" Dec 05 10:37:57.932741 master-0 kubenswrapper[10444]: E1205 10:37:57.932483 10444 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Dec 05 10:37:57.932741 master-0 kubenswrapper[10444]: E1205 10:37:57.932526 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls podName:58baad85-de54-49e7-a13e-d470d9c50d11 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:58.432514062 +0000 UTC m=+19.024426629 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls") pod "dns-default-4vxng" (UID: "58baad85-de54-49e7-a13e-d470d9c50d11") : secret "dns-default-metrics-tls" not found Dec 05 10:37:57.965079 master-0 kubenswrapper[10444]: I1205 10:37:57.965038 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx7jr\" (UniqueName: \"kubernetes.io/projected/58baad85-de54-49e7-a13e-d470d9c50d11-kube-api-access-jx7jr\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:57.971893 master-0 kubenswrapper[10444]: I1205 10:37:57.971852 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" event={"ID":"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082","Type":"ContainerStarted","Data":"0e02576ddbd856f2a1c30d4bbdcb0b27384c73dfe012be69bbb5c2168f8df7fd"} Dec 05 10:37:57.973660 master-0 kubenswrapper[10444]: I1205 10:37:57.973173 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" event={"ID":"8d76404b-6d62-4a61-b6f6-0c8073eba198","Type":"ContainerStarted","Data":"e9475cd437d6e0adb87102a00093bc37d7716f625ab1f62c89982010f22e3b2c"} Dec 05 10:37:57.975709 master-0 kubenswrapper[10444]: I1205 10:37:57.974642 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" event={"ID":"8c649a16-c187-412e-b5da-62a00bee38ab","Type":"ContainerStarted","Data":"f5a80fe6f7181cedae59e133ffaba19310d890b95762eab50ea91cd1fe447ad3"} Dec 05 10:37:57.975709 master-0 kubenswrapper[10444]: I1205 10:37:57.974673 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" event={"ID":"8c649a16-c187-412e-b5da-62a00bee38ab","Type":"ContainerStarted","Data":"77b83e99ece1e07058ac092ebb18d76d85bb5a90292440cd154e967f63253776"} Dec 05 10:37:57.975977 master-0 kubenswrapper[10444]: I1205 10:37:57.975950 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8gjgm" event={"ID":"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c","Type":"ContainerStarted","Data":"62e34bf231128449a630a3dad98495dd55715547e65586a815afe200566d05a0"} Dec 05 10:37:57.978783 master-0 kubenswrapper[10444]: I1205 10:37:57.978718 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" event={"ID":"49051e6e-5a2f-45c8-bad0-374514a91c07","Type":"ContainerStarted","Data":"0e525fa61aef33df1f1b067f329fadc308afa7e4b351f661375960287408ec4e"} Dec 05 10:37:57.980946 master-0 kubenswrapper[10444]: I1205 10:37:57.980902 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-1-master-0" event={"ID":"097d0e8f-46d4-4dfe-862f-54b2b9b21a89","Type":"ContainerStarted","Data":"f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c"} Dec 05 10:37:57.980946 master-0 kubenswrapper[10444]: I1205 10:37:57.980939 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-1-master-0" event={"ID":"097d0e8f-46d4-4dfe-862f-54b2b9b21a89","Type":"ContainerStarted","Data":"a16f63e21ba69da6613675e95c4ec003be827b3b94b2ba8c9b38f86425a10d28"} Dec 05 10:37:57.985962 master-0 kubenswrapper[10444]: I1205 10:37:57.985881 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" event={"ID":"926263c4-ec5b-41cb-9c30-0c88f636035f","Type":"ContainerStarted","Data":"ba6a43bb61d242c8e032f7cab8a9cf0e806cdf03c0a9ce44ec6c444a2216f39a"} Dec 05 10:37:57.987596 master-0 kubenswrapper[10444]: I1205 10:37:57.987551 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" event={"ID":"f7b29f89-e42d-4e53-ad14-05efdce933f0","Type":"ContainerStarted","Data":"578f760602feb29ac50a695c57a9184e5557511fef0ad085d002343f883452ae"} Dec 05 10:37:57.989920 master-0 kubenswrapper[10444]: I1205 10:37:57.989882 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" event={"ID":"22676fac-b770-4937-9bee-7478bd1babb7","Type":"ContainerStarted","Data":"7f5edea6bac7d86c19774c19407fa60083727d4b313c43d3272af6d979af6367"} Dec 05 10:37:57.990019 master-0 kubenswrapper[10444]: I1205 10:37:57.989926 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" event={"ID":"22676fac-b770-4937-9bee-7478bd1babb7","Type":"ContainerStarted","Data":"fd72b02943eacc00b1f17176da63bcc9ef38606a44cb5938204aea07936c36c7"} Dec 05 10:37:57.993397 master-0 kubenswrapper[10444]: I1205 10:37:57.992414 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" event={"ID":"5a1bdc70-6412-47e0-8330-04d796cc8d55","Type":"ContainerStarted","Data":"ec2b427f6d8abd4f8a13c4f08ae02d10372f1d8a84dd61d850f7c5ded0595e15"} Dec 05 10:37:57.993397 master-0 kubenswrapper[10444]: I1205 10:37:57.992753 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" event={"ID":"5a1bdc70-6412-47e0-8330-04d796cc8d55","Type":"ContainerStarted","Data":"3d73cb25f465febadbba348276b412b395b7d2893236bc1df51daa50a77648d5"} Dec 05 10:37:57.997645 master-0 kubenswrapper[10444]: I1205 10:37:57.997598 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" event={"ID":"a722cda9-29a0-4b7f-8e1d-9a8950ed765a","Type":"ContainerStarted","Data":"aa871dfa794fea5d166618a43475fe935ef7d4d3799afbc3d92debffe4b76081"} Dec 05 10:37:58.033850 master-0 kubenswrapper[10444]: I1205 10:37:58.032757 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kube-api-access\") pod \"installer-1-master-0\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " pod="openshift-etcd/installer-1-master-0" Dec 05 10:37:58.033850 master-0 kubenswrapper[10444]: I1205 10:37:58.032805 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " pod="openshift-etcd/installer-1-master-0" Dec 05 10:37:58.033850 master-0 kubenswrapper[10444]: I1205 10:37:58.032845 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-var-lock\") pod \"installer-1-master-0\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " pod="openshift-etcd/installer-1-master-0" Dec 05 10:37:58.033850 master-0 kubenswrapper[10444]: I1205 10:37:58.032914 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-var-lock\") pod \"installer-1-master-0\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " pod="openshift-etcd/installer-1-master-0" Dec 05 10:37:58.038750 master-0 kubenswrapper[10444]: I1205 10:37:58.035929 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " pod="openshift-etcd/installer-1-master-0" Dec 05 10:37:58.055184 master-0 kubenswrapper[10444]: I1205 10:37:58.054218 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/installer-1-master-0" podStartSLOduration=4.054198402 podStartE2EDuration="4.054198402s" podCreationTimestamp="2025-12-05 10:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:37:58.051371822 +0000 UTC m=+18.643284379" watchObservedRunningTime="2025-12-05 10:37:58.054198402 +0000 UTC m=+18.646110969" Dec 05 10:37:58.067816 master-0 kubenswrapper[10444]: I1205 10:37:58.065509 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kube-api-access\") pod \"installer-1-master-0\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " pod="openshift-etcd/installer-1-master-0" Dec 05 10:37:58.127199 master-0 kubenswrapper[10444]: I1205 10:37:58.127086 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-qkccw"] Dec 05 10:37:58.134330 master-0 kubenswrapper[10444]: I1205 10:37:58.134287 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qkccw" Dec 05 10:37:58.199644 master-0 kubenswrapper[10444]: I1205 10:37:58.199590 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-1-master-0" Dec 05 10:37:58.237292 master-0 kubenswrapper[10444]: I1205 10:37:58.237001 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/baee05cb-62ad-4eda-8a13-e317eae840e5-hosts-file\") pod \"node-resolver-qkccw\" (UID: \"baee05cb-62ad-4eda-8a13-e317eae840e5\") " pod="openshift-dns/node-resolver-qkccw" Dec 05 10:37:58.237292 master-0 kubenswrapper[10444]: I1205 10:37:58.237095 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-564nt\" (UniqueName: \"kubernetes.io/projected/baee05cb-62ad-4eda-8a13-e317eae840e5-kube-api-access-564nt\") pod \"node-resolver-qkccw\" (UID: \"baee05cb-62ad-4eda-8a13-e317eae840e5\") " pod="openshift-dns/node-resolver-qkccw" Dec 05 10:37:58.338977 master-0 kubenswrapper[10444]: I1205 10:37:58.338731 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/baee05cb-62ad-4eda-8a13-e317eae840e5-hosts-file\") pod \"node-resolver-qkccw\" (UID: \"baee05cb-62ad-4eda-8a13-e317eae840e5\") " pod="openshift-dns/node-resolver-qkccw" Dec 05 10:37:58.338977 master-0 kubenswrapper[10444]: I1205 10:37:58.338794 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-564nt\" (UniqueName: \"kubernetes.io/projected/baee05cb-62ad-4eda-8a13-e317eae840e5-kube-api-access-564nt\") pod \"node-resolver-qkccw\" (UID: \"baee05cb-62ad-4eda-8a13-e317eae840e5\") " pod="openshift-dns/node-resolver-qkccw" Dec 05 10:37:58.338977 master-0 kubenswrapper[10444]: I1205 10:37:58.338861 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/baee05cb-62ad-4eda-8a13-e317eae840e5-hosts-file\") pod \"node-resolver-qkccw\" (UID: \"baee05cb-62ad-4eda-8a13-e317eae840e5\") " pod="openshift-dns/node-resolver-qkccw" Dec 05 10:37:58.361306 master-0 kubenswrapper[10444]: I1205 10:37:58.361255 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-564nt\" (UniqueName: \"kubernetes.io/projected/baee05cb-62ad-4eda-8a13-e317eae840e5-kube-api-access-564nt\") pod \"node-resolver-qkccw\" (UID: \"baee05cb-62ad-4eda-8a13-e317eae840e5\") " pod="openshift-dns/node-resolver-qkccw" Dec 05 10:37:58.440116 master-0 kubenswrapper[10444]: I1205 10:37:58.440063 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:58.440515 master-0 kubenswrapper[10444]: E1205 10:37:58.440219 10444 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: secret "dns-default-metrics-tls" not found Dec 05 10:37:58.440515 master-0 kubenswrapper[10444]: E1205 10:37:58.440275 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls podName:58baad85-de54-49e7-a13e-d470d9c50d11 nodeName:}" failed. No retries permitted until 2025-12-05 10:37:59.440259636 +0000 UTC m=+20.032172203 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls") pod "dns-default-4vxng" (UID: "58baad85-de54-49e7-a13e-d470d9c50d11") : secret "dns-default-metrics-tls" not found Dec 05 10:37:58.465859 master-0 kubenswrapper[10444]: I1205 10:37:58.465805 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-qkccw" Dec 05 10:37:58.487483 master-0 kubenswrapper[10444]: W1205 10:37:58.487413 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbaee05cb_62ad_4eda_8a13_e317eae840e5.slice/crio-fec7b5fa2a30a341bff8e373ec4fbcd19900103be9e9af536060bcabaccdcce1 WatchSource:0}: Error finding container fec7b5fa2a30a341bff8e373ec4fbcd19900103be9e9af536060bcabaccdcce1: Status 404 returned error can't find the container with id fec7b5fa2a30a341bff8e373ec4fbcd19900103be9e9af536060bcabaccdcce1 Dec 05 10:37:58.644791 master-0 kubenswrapper[10444]: I1205 10:37:58.644694 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:58.652480 master-0 kubenswrapper[10444]: I1205 10:37:58.652346 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:58.663502 master-0 kubenswrapper[10444]: I1205 10:37:58.663451 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd/installer-1-master-0"] Dec 05 10:37:58.674238 master-0 kubenswrapper[10444]: W1205 10:37:58.674175 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode149cbe7_1ec3_4a06_af61_3a8906b8e9ef.slice/crio-a87c5cb2ce0ea32814514a0a2632b27b31a950e85396e1a0393057f93ec2e154 WatchSource:0}: Error finding container a87c5cb2ce0ea32814514a0a2632b27b31a950e85396e1a0393057f93ec2e154: Status 404 returned error can't find the container with id a87c5cb2ce0ea32814514a0a2632b27b31a950e85396e1a0393057f93ec2e154 Dec 05 10:37:58.825442 master-0 kubenswrapper[10444]: I1205 10:37:58.822632 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:37:59.014261 master-0 kubenswrapper[10444]: I1205 10:37:59.013655 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qkccw" event={"ID":"baee05cb-62ad-4eda-8a13-e317eae840e5","Type":"ContainerStarted","Data":"803c199236437bf9c0abc7fd5dc767458ec419cf17d83c9625161d425d0c0453"} Dec 05 10:37:59.014261 master-0 kubenswrapper[10444]: I1205 10:37:59.013731 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-qkccw" event={"ID":"baee05cb-62ad-4eda-8a13-e317eae840e5","Type":"ContainerStarted","Data":"fec7b5fa2a30a341bff8e373ec4fbcd19900103be9e9af536060bcabaccdcce1"} Dec 05 10:37:59.016877 master-0 kubenswrapper[10444]: I1205 10:37:59.016849 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-1-master-0" event={"ID":"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef","Type":"ContainerStarted","Data":"a87c5cb2ce0ea32814514a0a2632b27b31a950e85396e1a0393057f93ec2e154"} Dec 05 10:37:59.038007 master-0 kubenswrapper[10444]: I1205 10:37:59.037941 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-qkccw" podStartSLOduration=1.037920283 podStartE2EDuration="1.037920283s" podCreationTimestamp="2025-12-05 10:37:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:37:59.0360407 +0000 UTC m=+19.627953287" watchObservedRunningTime="2025-12-05 10:37:59.037920283 +0000 UTC m=+19.629832850" Dec 05 10:37:59.148578 master-0 kubenswrapper[10444]: I1205 10:37:59.145771 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6"] Dec 05 10:37:59.148578 master-0 kubenswrapper[10444]: E1205 10:37:59.146112 10444 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" podUID="61fd412b-c1c2-4902-9564-aab7a8b17845" Dec 05 10:37:59.157468 master-0 kubenswrapper[10444]: I1205 10:37:59.157236 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:59.159688 master-0 kubenswrapper[10444]: I1205 10:37:59.159222 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca\") pod \"controller-manager-6d9cb7b7fc-f9nz6\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:37:59.218628 master-0 kubenswrapper[10444]: I1205 10:37:59.218551 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr"] Dec 05 10:37:59.219065 master-0 kubenswrapper[10444]: E1205 10:37:59.218833 10444 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" podUID="98fd575c-ffc4-44c2-9947-c18d3f53e2c9" Dec 05 10:37:59.459511 master-0 kubenswrapper[10444]: I1205 10:37:59.459401 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:59.464730 master-0 kubenswrapper[10444]: I1205 10:37:59.464672 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:37:59.495120 master-0 kubenswrapper[10444]: I1205 10:37:59.494869 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-4vxng" Dec 05 10:38:00.022352 master-0 kubenswrapper[10444]: I1205 10:38:00.022288 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:38:00.023202 master-0 kubenswrapper[10444]: I1205 10:38:00.022293 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:38:00.030788 master-0 kubenswrapper[10444]: I1205 10:38:00.030736 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:38:00.036541 master-0 kubenswrapper[10444]: I1205 10:38:00.036459 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:38:00.165341 master-0 kubenswrapper[10444]: I1205 10:38:00.165274 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwpcs\" (UniqueName: \"kubernetes.io/projected/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-kube-api-access-fwpcs\") pod \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " Dec 05 10:38:00.165341 master-0 kubenswrapper[10444]: I1205 10:38:00.165336 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca\") pod \"61fd412b-c1c2-4902-9564-aab7a8b17845\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " Dec 05 10:38:00.165777 master-0 kubenswrapper[10444]: I1205 10:38:00.165374 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfh6j\" (UniqueName: \"kubernetes.io/projected/61fd412b-c1c2-4902-9564-aab7a8b17845-kube-api-access-tfh6j\") pod \"61fd412b-c1c2-4902-9564-aab7a8b17845\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " Dec 05 10:38:00.165777 master-0 kubenswrapper[10444]: I1205 10:38:00.165456 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-config\") pod \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " Dec 05 10:38:00.165777 master-0 kubenswrapper[10444]: I1205 10:38:00.165621 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert\") pod \"61fd412b-c1c2-4902-9564-aab7a8b17845\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " Dec 05 10:38:00.165777 master-0 kubenswrapper[10444]: I1205 10:38:00.165675 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-config\") pod \"61fd412b-c1c2-4902-9564-aab7a8b17845\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " Dec 05 10:38:00.165777 master-0 kubenswrapper[10444]: I1205 10:38:00.165700 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-proxy-ca-bundles\") pod \"61fd412b-c1c2-4902-9564-aab7a8b17845\" (UID: \"61fd412b-c1c2-4902-9564-aab7a8b17845\") " Dec 05 10:38:00.166033 master-0 kubenswrapper[10444]: I1205 10:38:00.165780 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert\") pod \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\" (UID: \"98fd575c-ffc4-44c2-9947-c18d3f53e2c9\") " Dec 05 10:38:00.166033 master-0 kubenswrapper[10444]: I1205 10:38:00.165904 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca" (OuterVolumeSpecName: "client-ca") pod "61fd412b-c1c2-4902-9564-aab7a8b17845" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:38:00.166222 master-0 kubenswrapper[10444]: I1205 10:38:00.166187 10444 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-client-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:00.166222 master-0 kubenswrapper[10444]: I1205 10:38:00.166201 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "61fd412b-c1c2-4902-9564-aab7a8b17845" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:38:00.166514 master-0 kubenswrapper[10444]: I1205 10:38:00.166262 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-config" (OuterVolumeSpecName: "config") pod "61fd412b-c1c2-4902-9564-aab7a8b17845" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:38:00.166569 master-0 kubenswrapper[10444]: I1205 10:38:00.166515 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-config" (OuterVolumeSpecName: "config") pod "98fd575c-ffc4-44c2-9947-c18d3f53e2c9" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:38:00.168740 master-0 kubenswrapper[10444]: I1205 10:38:00.168698 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-kube-api-access-fwpcs" (OuterVolumeSpecName: "kube-api-access-fwpcs") pod "98fd575c-ffc4-44c2-9947-c18d3f53e2c9" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9"). InnerVolumeSpecName "kube-api-access-fwpcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:38:00.168977 master-0 kubenswrapper[10444]: I1205 10:38:00.168842 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "98fd575c-ffc4-44c2-9947-c18d3f53e2c9" (UID: "98fd575c-ffc4-44c2-9947-c18d3f53e2c9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:38:00.169253 master-0 kubenswrapper[10444]: I1205 10:38:00.169221 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "61fd412b-c1c2-4902-9564-aab7a8b17845" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:38:00.170737 master-0 kubenswrapper[10444]: I1205 10:38:00.170673 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61fd412b-c1c2-4902-9564-aab7a8b17845-kube-api-access-tfh6j" (OuterVolumeSpecName: "kube-api-access-tfh6j") pod "61fd412b-c1c2-4902-9564-aab7a8b17845" (UID: "61fd412b-c1c2-4902-9564-aab7a8b17845"). InnerVolumeSpecName "kube-api-access-tfh6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:38:00.266661 master-0 kubenswrapper[10444]: I1205 10:38:00.266586 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwpcs\" (UniqueName: \"kubernetes.io/projected/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-kube-api-access-fwpcs\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:00.266661 master-0 kubenswrapper[10444]: I1205 10:38:00.266629 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfh6j\" (UniqueName: \"kubernetes.io/projected/61fd412b-c1c2-4902-9564-aab7a8b17845-kube-api-access-tfh6j\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:00.266661 master-0 kubenswrapper[10444]: I1205 10:38:00.266639 10444 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:00.266661 master-0 kubenswrapper[10444]: I1205 10:38:00.266650 10444 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61fd412b-c1c2-4902-9564-aab7a8b17845-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:00.266661 master-0 kubenswrapper[10444]: I1205 10:38:00.266660 10444 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:00.266661 master-0 kubenswrapper[10444]: I1205 10:38:00.266668 10444 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/61fd412b-c1c2-4902-9564-aab7a8b17845-proxy-ca-bundles\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:00.266661 master-0 kubenswrapper[10444]: I1205 10:38:00.266677 10444 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:01.028671 master-0 kubenswrapper[10444]: I1205 10:38:01.026715 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr" Dec 05 10:38:01.028671 master-0 kubenswrapper[10444]: I1205 10:38:01.026716 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6" Dec 05 10:38:01.084650 master-0 kubenswrapper[10444]: I1205 10:38:01.084591 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6458c74b4c-4gvlc"] Dec 05 10:38:01.085316 master-0 kubenswrapper[10444]: I1205 10:38:01.085287 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.087591 master-0 kubenswrapper[10444]: I1205 10:38:01.087537 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 10:38:01.087716 master-0 kubenswrapper[10444]: I1205 10:38:01.087654 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 10:38:01.088092 master-0 kubenswrapper[10444]: I1205 10:38:01.088061 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 10:38:01.091661 master-0 kubenswrapper[10444]: I1205 10:38:01.091622 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 10:38:01.092729 master-0 kubenswrapper[10444]: I1205 10:38:01.092697 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 10:38:01.093647 master-0 kubenswrapper[10444]: I1205 10:38:01.093565 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6"] Dec 05 10:38:01.101264 master-0 kubenswrapper[10444]: I1205 10:38:01.098594 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 10:38:01.107583 master-0 kubenswrapper[10444]: I1205 10:38:01.107458 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6458c74b4c-4gvlc"] Dec 05 10:38:01.108638 master-0 kubenswrapper[10444]: I1205 10:38:01.108585 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6d9cb7b7fc-f9nz6"] Dec 05 10:38:01.141711 master-0 kubenswrapper[10444]: I1205 10:38:01.141648 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr"] Dec 05 10:38:01.143086 master-0 kubenswrapper[10444]: I1205 10:38:01.143041 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76d4564964-xm2tr"] Dec 05 10:38:01.188793 master-0 kubenswrapper[10444]: I1205 10:38:01.188745 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-config\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.188963 master-0 kubenswrapper[10444]: I1205 10:38:01.188809 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/941ce297-c739-41e4-804e-fd7d83726170-serving-cert\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.188963 master-0 kubenswrapper[10444]: I1205 10:38:01.188846 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-client-ca\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.188963 master-0 kubenswrapper[10444]: I1205 10:38:01.188903 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-proxy-ca-bundles\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.189333 master-0 kubenswrapper[10444]: I1205 10:38:01.189033 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnmjt\" (UniqueName: \"kubernetes.io/projected/941ce297-c739-41e4-804e-fd7d83726170-kube-api-access-bnmjt\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.189333 master-0 kubenswrapper[10444]: I1205 10:38:01.189133 10444 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/98fd575c-ffc4-44c2-9947-c18d3f53e2c9-client-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:01.293560 master-0 kubenswrapper[10444]: I1205 10:38:01.291630 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-config\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.293560 master-0 kubenswrapper[10444]: I1205 10:38:01.291695 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/941ce297-c739-41e4-804e-fd7d83726170-serving-cert\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.293560 master-0 kubenswrapper[10444]: I1205 10:38:01.291724 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-client-ca\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.293560 master-0 kubenswrapper[10444]: I1205 10:38:01.291796 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-proxy-ca-bundles\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.293560 master-0 kubenswrapper[10444]: I1205 10:38:01.291835 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnmjt\" (UniqueName: \"kubernetes.io/projected/941ce297-c739-41e4-804e-fd7d83726170-kube-api-access-bnmjt\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.293560 master-0 kubenswrapper[10444]: I1205 10:38:01.293250 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-client-ca\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.293560 master-0 kubenswrapper[10444]: I1205 10:38:01.293460 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-config\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.301036 master-0 kubenswrapper[10444]: I1205 10:38:01.297848 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-proxy-ca-bundles\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.301036 master-0 kubenswrapper[10444]: I1205 10:38:01.300120 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/941ce297-c739-41e4-804e-fd7d83726170-serving-cert\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.314327 master-0 kubenswrapper[10444]: I1205 10:38:01.314285 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnmjt\" (UniqueName: \"kubernetes.io/projected/941ce297-c739-41e4-804e-fd7d83726170-kube-api-access-bnmjt\") pod \"controller-manager-6458c74b4c-4gvlc\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.412781 master-0 kubenswrapper[10444]: I1205 10:38:01.412568 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:01.606363 master-0 kubenswrapper[10444]: I1205 10:38:01.606306 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61fd412b-c1c2-4902-9564-aab7a8b17845" path="/var/lib/kubelet/pods/61fd412b-c1c2-4902-9564-aab7a8b17845/volumes" Dec 05 10:38:01.606836 master-0 kubenswrapper[10444]: I1205 10:38:01.606776 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98fd575c-ffc4-44c2-9947-c18d3f53e2c9" path="/var/lib/kubelet/pods/98fd575c-ffc4-44c2-9947-c18d3f53e2c9/volumes" Dec 05 10:38:01.642297 master-0 kubenswrapper[10444]: I1205 10:38:01.640326 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-5b9fd577f8-6sxcx"] Dec 05 10:38:01.695309 master-0 kubenswrapper[10444]: I1205 10:38:01.694397 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-4vxng"] Dec 05 10:38:01.708451 master-0 kubenswrapper[10444]: W1205 10:38:01.708361 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58baad85_de54_49e7_a13e_d470d9c50d11.slice/crio-487c60a3656ecce96053a1bf83cb0f4b361d27909df0d101db089d62b7524b7b WatchSource:0}: Error finding container 487c60a3656ecce96053a1bf83cb0f4b361d27909df0d101db089d62b7524b7b: Status 404 returned error can't find the container with id 487c60a3656ecce96053a1bf83cb0f4b361d27909df0d101db089d62b7524b7b Dec 05 10:38:01.932873 master-0 kubenswrapper[10444]: I1205 10:38:01.932572 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6458c74b4c-4gvlc"] Dec 05 10:38:02.036007 master-0 kubenswrapper[10444]: I1205 10:38:02.035927 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" event={"ID":"8d76404b-6d62-4a61-b6f6-0c8073eba198","Type":"ContainerStarted","Data":"d70a4de37cbf6d2709586eaecfa831636df1b7ccf32f015e3097c3c53b817cac"} Dec 05 10:38:02.036007 master-0 kubenswrapper[10444]: I1205 10:38:02.036000 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" event={"ID":"8d76404b-6d62-4a61-b6f6-0c8073eba198","Type":"ContainerStarted","Data":"163ef91a3ca09ff8a80da2c166f293609027eb6a5251a7136d985bf676587d9d"} Dec 05 10:38:02.037580 master-0 kubenswrapper[10444]: I1205 10:38:02.037531 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-1-master-0" event={"ID":"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef","Type":"ContainerStarted","Data":"86721e833ddfe5b80f087deb054f43ae1d6033b98f09d3f999eb1efccfb11215"} Dec 05 10:38:02.040035 master-0 kubenswrapper[10444]: I1205 10:38:02.039984 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4vxng" event={"ID":"58baad85-de54-49e7-a13e-d470d9c50d11","Type":"ContainerStarted","Data":"487c60a3656ecce96053a1bf83cb0f4b361d27909df0d101db089d62b7524b7b"} Dec 05 10:38:02.042239 master-0 kubenswrapper[10444]: I1205 10:38:02.041923 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8gjgm" event={"ID":"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c","Type":"ContainerStarted","Data":"e23d7aa3a66016c59bb88b79bb82a8734a8b41d6c4b402808c30e2f9a5f175a2"} Dec 05 10:38:02.042239 master-0 kubenswrapper[10444]: I1205 10:38:02.041951 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8gjgm" event={"ID":"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c","Type":"ContainerStarted","Data":"d32e5b11045f2e9dc6127afe0f5878a07b8c3b1ed39b689d816526629b2597df"} Dec 05 10:38:02.053604 master-0 kubenswrapper[10444]: I1205 10:38:02.043608 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" event={"ID":"a722cda9-29a0-4b7f-8e1d-9a8950ed765a","Type":"ContainerStarted","Data":"a45f2b264fecbb76e76a1090ae5db456fe685490af65388ce3ac4ed68eb080b2"} Dec 05 10:38:02.053604 master-0 kubenswrapper[10444]: I1205 10:38:02.052320 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" event={"ID":"926263c4-ec5b-41cb-9c30-0c88f636035f","Type":"ContainerStarted","Data":"3504f06691ba2b115136a2091ae95118383b7c2cc711e4550b1e60138be21510"} Dec 05 10:38:02.053604 master-0 kubenswrapper[10444]: I1205 10:38:02.053408 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:38:02.055252 master-0 kubenswrapper[10444]: I1205 10:38:02.055147 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" event={"ID":"941ce297-c739-41e4-804e-fd7d83726170","Type":"ContainerStarted","Data":"1e491ceac32476146ec752c9cf678575f2f44cfeb6d755985ac117ae34fbb8ac"} Dec 05 10:38:02.057409 master-0 kubenswrapper[10444]: I1205 10:38:02.057354 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" event={"ID":"d5198438-06ae-4e63-a7e3-950ba23bba9c","Type":"ContainerStarted","Data":"59a797eb008b1a85db4180d24d857e8f08956acdd19cbd4d960f8adc3117a02b"} Dec 05 10:38:02.062033 master-0 kubenswrapper[10444]: I1205 10:38:02.061978 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:38:02.104047 master-0 kubenswrapper[10444]: I1205 10:38:02.097055 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/installer-1-master-0" podStartSLOduration=5.097036609 podStartE2EDuration="5.097036609s" podCreationTimestamp="2025-12-05 10:37:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:02.09635072 +0000 UTC m=+22.688263307" watchObservedRunningTime="2025-12-05 10:38:02.097036609 +0000 UTC m=+22.688949166" Dec 05 10:38:02.369888 master-0 kubenswrapper[10444]: I1205 10:38:02.369606 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-85b8f855df-8g52w"] Dec 05 10:38:02.381151 master-0 kubenswrapper[10444]: I1205 10:38:02.380229 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.382540 master-0 kubenswrapper[10444]: I1205 10:38:02.382432 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 10:38:02.382754 master-0 kubenswrapper[10444]: I1205 10:38:02.382734 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 10:38:02.382906 master-0 kubenswrapper[10444]: I1205 10:38:02.382888 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 10:38:02.383350 master-0 kubenswrapper[10444]: I1205 10:38:02.383300 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 10:38:02.385001 master-0 kubenswrapper[10444]: I1205 10:38:02.384233 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 10:38:02.385001 master-0 kubenswrapper[10444]: I1205 10:38:02.384414 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 10:38:02.386478 master-0 kubenswrapper[10444]: I1205 10:38:02.386238 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 10:38:02.388463 master-0 kubenswrapper[10444]: I1205 10:38:02.388405 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-85b8f855df-8g52w"] Dec 05 10:38:02.388869 master-0 kubenswrapper[10444]: I1205 10:38:02.388827 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 10:38:02.407749 master-0 kubenswrapper[10444]: I1205 10:38:02.407698 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-serving-ca\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.407891 master-0 kubenswrapper[10444]: I1205 10:38:02.407760 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-policies\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.407891 master-0 kubenswrapper[10444]: I1205 10:38:02.407780 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-encryption-config\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.407891 master-0 kubenswrapper[10444]: I1205 10:38:02.407797 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltjt6\" (UniqueName: \"kubernetes.io/projected/ca45c52e-fb30-4e7c-8c3f-e685c0909916-kube-api-access-ltjt6\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.407891 master-0 kubenswrapper[10444]: I1205 10:38:02.407814 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-trusted-ca-bundle\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.407891 master-0 kubenswrapper[10444]: I1205 10:38:02.407830 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-dir\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.407891 master-0 kubenswrapper[10444]: I1205 10:38:02.407850 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-client\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.407891 master-0 kubenswrapper[10444]: I1205 10:38:02.407874 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-serving-cert\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510042 master-0 kubenswrapper[10444]: I1205 10:38:02.508520 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-serving-ca\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510042 master-0 kubenswrapper[10444]: I1205 10:38:02.508587 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-policies\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510042 master-0 kubenswrapper[10444]: I1205 10:38:02.508604 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-encryption-config\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510042 master-0 kubenswrapper[10444]: I1205 10:38:02.508625 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltjt6\" (UniqueName: \"kubernetes.io/projected/ca45c52e-fb30-4e7c-8c3f-e685c0909916-kube-api-access-ltjt6\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510042 master-0 kubenswrapper[10444]: I1205 10:38:02.508645 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-trusted-ca-bundle\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510042 master-0 kubenswrapper[10444]: I1205 10:38:02.509282 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-dir\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510042 master-0 kubenswrapper[10444]: I1205 10:38:02.509306 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-client\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510042 master-0 kubenswrapper[10444]: I1205 10:38:02.509325 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-serving-cert\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510042 master-0 kubenswrapper[10444]: I1205 10:38:02.509504 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-dir\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510042 master-0 kubenswrapper[10444]: I1205 10:38:02.509567 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-trusted-ca-bundle\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510042 master-0 kubenswrapper[10444]: I1205 10:38:02.509960 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-policies\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.510899 master-0 kubenswrapper[10444]: I1205 10:38:02.510270 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-serving-ca\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.527703 master-0 kubenswrapper[10444]: I1205 10:38:02.514165 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-encryption-config\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.527703 master-0 kubenswrapper[10444]: I1205 10:38:02.514208 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-serving-cert\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:02.527703 master-0 kubenswrapper[10444]: I1205 10:38:02.514277 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-client\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:05.907844 master-0 kubenswrapper[10444]: I1205 10:38:05.907788 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln"] Dec 05 10:38:05.909278 master-0 kubenswrapper[10444]: I1205 10:38:05.909244 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:05.918130 master-0 kubenswrapper[10444]: I1205 10:38:05.918065 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 10:38:05.928974 master-0 kubenswrapper[10444]: I1205 10:38:05.924838 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 10:38:05.928974 master-0 kubenswrapper[10444]: I1205 10:38:05.925204 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 10:38:05.928974 master-0 kubenswrapper[10444]: I1205 10:38:05.925489 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 10:38:05.928974 master-0 kubenswrapper[10444]: I1205 10:38:05.925697 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 10:38:05.948451 master-0 kubenswrapper[10444]: I1205 10:38:05.934600 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltjt6\" (UniqueName: \"kubernetes.io/projected/ca45c52e-fb30-4e7c-8c3f-e685c0909916-kube-api-access-ltjt6\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:05.967666 master-0 kubenswrapper[10444]: I1205 10:38:05.967495 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2118a526-c5ce-4c64-b955-48629a5db35b-serving-cert\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:05.967666 master-0 kubenswrapper[10444]: I1205 10:38:05.967570 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbl76\" (UniqueName: \"kubernetes.io/projected/2118a526-c5ce-4c64-b955-48629a5db35b-kube-api-access-rbl76\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:05.967666 master-0 kubenswrapper[10444]: I1205 10:38:05.967609 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-config\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:05.967666 master-0 kubenswrapper[10444]: I1205 10:38:05.967636 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-client-ca\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:06.012192 master-0 kubenswrapper[10444]: I1205 10:38:06.012136 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:06.069077 master-0 kubenswrapper[10444]: I1205 10:38:06.068996 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbl76\" (UniqueName: \"kubernetes.io/projected/2118a526-c5ce-4c64-b955-48629a5db35b-kube-api-access-rbl76\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:06.069077 master-0 kubenswrapper[10444]: I1205 10:38:06.069068 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-config\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:06.069387 master-0 kubenswrapper[10444]: I1205 10:38:06.069102 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-client-ca\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:06.069387 master-0 kubenswrapper[10444]: I1205 10:38:06.069166 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2118a526-c5ce-4c64-b955-48629a5db35b-serving-cert\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:06.070483 master-0 kubenswrapper[10444]: I1205 10:38:06.070440 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-client-ca\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:06.072590 master-0 kubenswrapper[10444]: I1205 10:38:06.071700 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-config\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:06.072590 master-0 kubenswrapper[10444]: I1205 10:38:06.072227 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2118a526-c5ce-4c64-b955-48629a5db35b-serving-cert\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:07.733465 master-0 kubenswrapper[10444]: I1205 10:38:07.730757 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln"] Dec 05 10:38:07.951457 master-0 kubenswrapper[10444]: I1205 10:38:07.951378 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbl76\" (UniqueName: \"kubernetes.io/projected/2118a526-c5ce-4c64-b955-48629a5db35b-kube-api-access-rbl76\") pod \"route-controller-manager-7f6f96665d-4nkln\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:07.998533 master-0 kubenswrapper[10444]: I1205 10:38:07.995810 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-scheduler/installer-1-master-0"] Dec 05 10:38:07.998533 master-0 kubenswrapper[10444]: I1205 10:38:07.996039 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-scheduler/installer-1-master-0" podUID="097d0e8f-46d4-4dfe-862f-54b2b9b21a89" containerName="installer" containerID="cri-o://f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c" gracePeriod=30 Dec 05 10:38:08.078014 master-0 kubenswrapper[10444]: I1205 10:38:08.077952 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:08.570346 master-0 kubenswrapper[10444]: I1205 10:38:08.570289 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87"] Dec 05 10:38:08.571073 master-0 kubenswrapper[10444]: I1205 10:38:08.571043 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.578343 master-0 kubenswrapper[10444]: I1205 10:38:08.576636 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-catalogd"/"openshift-service-ca.crt" Dec 05 10:38:08.578343 master-0 kubenswrapper[10444]: I1205 10:38:08.576636 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-catalogd"/"kube-root-ca.crt" Dec 05 10:38:08.578343 master-0 kubenswrapper[10444]: I1205 10:38:08.577332 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-catalogd"/"catalogserver-cert" Dec 05 10:38:08.584595 master-0 kubenswrapper[10444]: I1205 10:38:08.584556 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-catalogd"/"catalogd-trusted-ca-bundle" Dec 05 10:38:08.589276 master-0 kubenswrapper[10444]: I1205 10:38:08.588377 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87"] Dec 05 10:38:08.614443 master-0 kubenswrapper[10444]: I1205 10:38:08.612332 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-ca-certs\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.614443 master-0 kubenswrapper[10444]: I1205 10:38:08.612398 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalogserver-certs\" (UniqueName: \"kubernetes.io/secret/a4fd453c-e667-4bdc-aa9e-3d95ff707200-catalogserver-certs\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.614443 master-0 kubenswrapper[10444]: I1205 10:38:08.612458 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-docker\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.614443 master-0 kubenswrapper[10444]: I1205 10:38:08.612495 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a4fd453c-e667-4bdc-aa9e-3d95ff707200-cache\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.614443 master-0 kubenswrapper[10444]: I1205 10:38:08.612532 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-containers\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.614443 master-0 kubenswrapper[10444]: I1205 10:38:08.612577 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmhhw\" (UniqueName: \"kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-kube-api-access-kmhhw\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.660027 master-0 kubenswrapper[10444]: I1205 10:38:08.651875 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt"] Dec 05 10:38:08.660027 master-0 kubenswrapper[10444]: I1205 10:38:08.652727 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.660027 master-0 kubenswrapper[10444]: I1205 10:38:08.654577 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt"] Dec 05 10:38:08.660027 master-0 kubenswrapper[10444]: I1205 10:38:08.655869 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-controller"/"openshift-service-ca.crt" Dec 05 10:38:08.660027 master-0 kubenswrapper[10444]: I1205 10:38:08.655869 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-controller"/"operator-controller-trusted-ca-bundle" Dec 05 10:38:08.660027 master-0 kubenswrapper[10444]: I1205 10:38:08.656544 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-controller"/"kube-root-ca.crt" Dec 05 10:38:08.713349 master-0 kubenswrapper[10444]: I1205 10:38:08.713285 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-ca-certs\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.713669 master-0 kubenswrapper[10444]: I1205 10:38:08.713616 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalogserver-certs\" (UniqueName: \"kubernetes.io/secret/a4fd453c-e667-4bdc-aa9e-3d95ff707200-catalogserver-certs\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.713730 master-0 kubenswrapper[10444]: I1205 10:38:08.713704 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-docker\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.713778 master-0 kubenswrapper[10444]: I1205 10:38:08.713761 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a4fd453c-e667-4bdc-aa9e-3d95ff707200-cache\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.713881 master-0 kubenswrapper[10444]: I1205 10:38:08.713838 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-containers\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.713937 master-0 kubenswrapper[10444]: I1205 10:38:08.713881 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-docker\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.713937 master-0 kubenswrapper[10444]: I1205 10:38:08.713916 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-containers\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.713937 master-0 kubenswrapper[10444]: I1205 10:38:08.713924 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmhhw\" (UniqueName: \"kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-kube-api-access-kmhhw\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.714153 master-0 kubenswrapper[10444]: I1205 10:38:08.714128 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a4fd453c-e667-4bdc-aa9e-3d95ff707200-cache\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.716582 master-0 kubenswrapper[10444]: I1205 10:38:08.716543 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalogserver-certs\" (UniqueName: \"kubernetes.io/secret/a4fd453c-e667-4bdc-aa9e-3d95ff707200-catalogserver-certs\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.716952 master-0 kubenswrapper[10444]: I1205 10:38:08.716919 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-ca-certs\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.772709 master-0 kubenswrapper[10444]: I1205 10:38:08.772658 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmhhw\" (UniqueName: \"kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-kube-api-access-kmhhw\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.814937 master-0 kubenswrapper[10444]: I1205 10:38:08.814883 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-ca-certs\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.815146 master-0 kubenswrapper[10444]: I1205 10:38:08.815000 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-containers\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.815146 master-0 kubenswrapper[10444]: I1205 10:38:08.815071 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-docker\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.815146 master-0 kubenswrapper[10444]: I1205 10:38:08.815090 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/592373ae-a952-4dd3-a9bc-f9c9c19c0802-cache\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.815146 master-0 kubenswrapper[10444]: I1205 10:38:08.815106 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd87p\" (UniqueName: \"kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-kube-api-access-rd87p\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.896155 master-0 kubenswrapper[10444]: I1205 10:38:08.895952 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:08.915511 master-0 kubenswrapper[10444]: I1205 10:38:08.915443 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-ca-certs\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.915511 master-0 kubenswrapper[10444]: I1205 10:38:08.915503 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-containers\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.915688 master-0 kubenswrapper[10444]: I1205 10:38:08.915597 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-docker\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.915688 master-0 kubenswrapper[10444]: I1205 10:38:08.915621 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/592373ae-a952-4dd3-a9bc-f9c9c19c0802-cache\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.915688 master-0 kubenswrapper[10444]: I1205 10:38:08.915647 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd87p\" (UniqueName: \"kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-kube-api-access-rd87p\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.915819 master-0 kubenswrapper[10444]: I1205 10:38:08.915755 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-containers\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.916202 master-0 kubenswrapper[10444]: I1205 10:38:08.916163 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/592373ae-a952-4dd3-a9bc-f9c9c19c0802-cache\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.916263 master-0 kubenswrapper[10444]: I1205 10:38:08.916217 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-docker\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:08.916310 master-0 kubenswrapper[10444]: E1205 10:38:08.916251 10444 projected.go:301] Couldn't get configMap payload openshift-operator-controller/operator-controller-trusted-ca-bundle: configmap references non-existent config key: ca-bundle.crt Dec 05 10:38:08.916310 master-0 kubenswrapper[10444]: E1205 10:38:08.916293 10444 projected.go:194] Error preparing data for projected volume ca-certs for pod openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt: configmap references non-existent config key: ca-bundle.crt Dec 05 10:38:08.916397 master-0 kubenswrapper[10444]: E1205 10:38:08.916382 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-ca-certs podName:592373ae-a952-4dd3-a9bc-f9c9c19c0802 nodeName:}" failed. No retries permitted until 2025-12-05 10:38:09.416357004 +0000 UTC m=+30.008269631 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ca-certs" (UniqueName: "kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-ca-certs") pod "operator-controller-controller-manager-7cbd59c7f8-dh5tt" (UID: "592373ae-a952-4dd3-a9bc-f9c9c19c0802") : configmap references non-existent config key: ca-bundle.crt Dec 05 10:38:08.951683 master-0 kubenswrapper[10444]: I1205 10:38:08.951565 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd87p\" (UniqueName: \"kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-kube-api-access-rd87p\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:09.420906 master-0 kubenswrapper[10444]: I1205 10:38:09.420841 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-ca-certs\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:09.424954 master-0 kubenswrapper[10444]: I1205 10:38:09.424910 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-ca-certs\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:09.582100 master-0 kubenswrapper[10444]: I1205 10:38:09.582030 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:10.737018 master-0 kubenswrapper[10444]: I1205 10:38:10.736947 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/installer-2-master-0"] Dec 05 10:38:10.737637 master-0 kubenswrapper[10444]: I1205 10:38:10.737600 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:10.751884 master-0 kubenswrapper[10444]: I1205 10:38:10.751793 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/installer-2-master-0"] Dec 05 10:38:10.940279 master-0 kubenswrapper[10444]: I1205 10:38:10.940185 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kube-api-access\") pod \"installer-2-master-0\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:10.940279 master-0 kubenswrapper[10444]: I1205 10:38:10.940240 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:10.940279 master-0 kubenswrapper[10444]: I1205 10:38:10.940285 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-var-lock\") pod \"installer-2-master-0\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:11.041099 master-0 kubenswrapper[10444]: I1205 10:38:11.041043 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kube-api-access\") pod \"installer-2-master-0\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:11.041286 master-0 kubenswrapper[10444]: I1205 10:38:11.041108 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:11.041286 master-0 kubenswrapper[10444]: I1205 10:38:11.041132 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-var-lock\") pod \"installer-2-master-0\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:11.041286 master-0 kubenswrapper[10444]: I1205 10:38:11.041209 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-var-lock\") pod \"installer-2-master-0\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:11.041286 master-0 kubenswrapper[10444]: I1205 10:38:11.041248 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:11.058705 master-0 kubenswrapper[10444]: I1205 10:38:11.058634 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kube-api-access\") pod \"installer-2-master-0\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:11.069449 master-0 kubenswrapper[10444]: I1205 10:38:11.069387 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:11.633211 master-0 kubenswrapper[10444]: I1205 10:38:11.633144 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:38:12.867853 master-0 kubenswrapper[10444]: I1205 10:38:12.867743 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m"] Dec 05 10:38:12.871364 master-0 kubenswrapper[10444]: I1205 10:38:12.868104 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" podUID="82ef99d4-41b7-4859-a5a1-4e31ce614a2a" containerName="cluster-version-operator" containerID="cri-o://8575f477d29594718db712e5d56ee7bf4ba0a8c565ae4540b71d0eb7f5d45356" gracePeriod=130 Dec 05 10:38:13.129908 master-0 kubenswrapper[10444]: I1205 10:38:13.129780 10444 generic.go:334] "Generic (PLEG): container finished" podID="82ef99d4-41b7-4859-a5a1-4e31ce614a2a" containerID="8575f477d29594718db712e5d56ee7bf4ba0a8c565ae4540b71d0eb7f5d45356" exitCode=0 Dec 05 10:38:13.129908 master-0 kubenswrapper[10444]: I1205 10:38:13.129836 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" event={"ID":"82ef99d4-41b7-4859-a5a1-4e31ce614a2a","Type":"ContainerDied","Data":"8575f477d29594718db712e5d56ee7bf4ba0a8c565ae4540b71d0eb7f5d45356"} Dec 05 10:38:13.450331 master-0 kubenswrapper[10444]: I1205 10:38:13.450279 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:38:13.575905 master-0 kubenswrapper[10444]: I1205 10:38:13.574803 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") pod \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " Dec 05 10:38:13.575905 master-0 kubenswrapper[10444]: I1205 10:38:13.574912 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-service-ca\") pod \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " Dec 05 10:38:13.575905 master-0 kubenswrapper[10444]: I1205 10:38:13.574958 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-kube-api-access\") pod \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " Dec 05 10:38:13.575905 master-0 kubenswrapper[10444]: I1205 10:38:13.575002 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-cvo-updatepayloads\") pod \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " Dec 05 10:38:13.575905 master-0 kubenswrapper[10444]: I1205 10:38:13.575080 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-ssl-certs\") pod \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\" (UID: \"82ef99d4-41b7-4859-a5a1-4e31ce614a2a\") " Dec 05 10:38:13.575905 master-0 kubenswrapper[10444]: I1205 10:38:13.575281 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-ssl-certs" (OuterVolumeSpecName: "etc-ssl-certs") pod "82ef99d4-41b7-4859-a5a1-4e31ce614a2a" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a"). InnerVolumeSpecName "etc-ssl-certs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:38:13.575905 master-0 kubenswrapper[10444]: I1205 10:38:13.575874 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-cvo-updatepayloads" (OuterVolumeSpecName: "etc-cvo-updatepayloads") pod "82ef99d4-41b7-4859-a5a1-4e31ce614a2a" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a"). InnerVolumeSpecName "etc-cvo-updatepayloads". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:38:13.576615 master-0 kubenswrapper[10444]: I1205 10:38:13.576555 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-service-ca" (OuterVolumeSpecName: "service-ca") pod "82ef99d4-41b7-4859-a5a1-4e31ce614a2a" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:38:13.587474 master-0 kubenswrapper[10444]: I1205 10:38:13.587152 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "82ef99d4-41b7-4859-a5a1-4e31ce614a2a" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:38:13.587474 master-0 kubenswrapper[10444]: I1205 10:38:13.587316 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "82ef99d4-41b7-4859-a5a1-4e31ce614a2a" (UID: "82ef99d4-41b7-4859-a5a1-4e31ce614a2a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:38:13.676693 master-0 kubenswrapper[10444]: I1205 10:38:13.676300 10444 reconciler_common.go:293] "Volume detached for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-cvo-updatepayloads\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:13.676693 master-0 kubenswrapper[10444]: I1205 10:38:13.676338 10444 reconciler_common.go:293] "Volume detached for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-etc-ssl-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:13.676693 master-0 kubenswrapper[10444]: I1205 10:38:13.676366 10444 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:13.676693 master-0 kubenswrapper[10444]: I1205 10:38:13.676381 10444 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-service-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:13.676693 master-0 kubenswrapper[10444]: I1205 10:38:13.676392 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/82ef99d4-41b7-4859-a5a1-4e31ce614a2a-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:13.719387 master-0 kubenswrapper[10444]: I1205 10:38:13.719334 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-85b8f855df-8g52w"] Dec 05 10:38:13.740533 master-0 kubenswrapper[10444]: W1205 10:38:13.740462 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca45c52e_fb30_4e7c_8c3f_e685c0909916.slice/crio-bd5095616b6a8494a79971c731d5ef49c3bae5cbd4111190f1e5dfde0e3d9554 WatchSource:0}: Error finding container bd5095616b6a8494a79971c731d5ef49c3bae5cbd4111190f1e5dfde0e3d9554: Status 404 returned error can't find the container with id bd5095616b6a8494a79971c731d5ef49c3bae5cbd4111190f1e5dfde0e3d9554 Dec 05 10:38:13.895974 master-0 kubenswrapper[10444]: I1205 10:38:13.894130 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt"] Dec 05 10:38:13.900211 master-0 kubenswrapper[10444]: W1205 10:38:13.900093 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod592373ae_a952_4dd3_a9bc_f9c9c19c0802.slice/crio-2c8d17056f01fe2b63601e7d3f2b689f6a1b5945d124ed1dd3ebcdd26090636d WatchSource:0}: Error finding container 2c8d17056f01fe2b63601e7d3f2b689f6a1b5945d124ed1dd3ebcdd26090636d: Status 404 returned error can't find the container with id 2c8d17056f01fe2b63601e7d3f2b689f6a1b5945d124ed1dd3ebcdd26090636d Dec 05 10:38:13.914407 master-0 kubenswrapper[10444]: I1205 10:38:13.914010 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/installer-2-master-0"] Dec 05 10:38:13.916165 master-0 kubenswrapper[10444]: I1205 10:38:13.916106 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87"] Dec 05 10:38:13.924377 master-0 kubenswrapper[10444]: W1205 10:38:13.924255 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9e7e6d3f_d579_4a5a_a0ad_bb4c6aa73a35.slice/crio-4751d52074fab20e20ae168d59e19aac68268e7927466f42d617e9b690cf2dab WatchSource:0}: Error finding container 4751d52074fab20e20ae168d59e19aac68268e7927466f42d617e9b690cf2dab: Status 404 returned error can't find the container with id 4751d52074fab20e20ae168d59e19aac68268e7927466f42d617e9b690cf2dab Dec 05 10:38:13.927015 master-0 kubenswrapper[10444]: W1205 10:38:13.926955 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4fd453c_e667_4bdc_aa9e_3d95ff707200.slice/crio-80b1488f509c166b5b116f44dccb9102304f263ee75bb4a43991cef27c102c84 WatchSource:0}: Error finding container 80b1488f509c166b5b116f44dccb9102304f263ee75bb4a43991cef27c102c84: Status 404 returned error can't find the container with id 80b1488f509c166b5b116f44dccb9102304f263ee75bb4a43991cef27c102c84 Dec 05 10:38:14.092697 master-0 kubenswrapper[10444]: I1205 10:38:14.090332 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln"] Dec 05 10:38:14.139627 master-0 kubenswrapper[10444]: I1205 10:38:14.138316 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" event={"ID":"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082","Type":"ContainerStarted","Data":"732d76d2aa1840bfa758b4121da0709dfcc66366de11514f51491dc82ad2dac3"} Dec 05 10:38:14.139627 master-0 kubenswrapper[10444]: I1205 10:38:14.139093 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:38:14.143448 master-0 kubenswrapper[10444]: I1205 10:38:14.141795 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4vxng" event={"ID":"58baad85-de54-49e7-a13e-d470d9c50d11","Type":"ContainerStarted","Data":"ac020d51da19428081ad1abf13deabd20a519ddb89317c7d1aa5a6b0778c818f"} Dec 05 10:38:14.143448 master-0 kubenswrapper[10444]: I1205 10:38:14.141846 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-4vxng" event={"ID":"58baad85-de54-49e7-a13e-d470d9c50d11","Type":"ContainerStarted","Data":"fe3f27a13d0a1b230b5d4febb56730fb5a507ed5a4564353c8ec784c56d4c692"} Dec 05 10:38:14.143448 master-0 kubenswrapper[10444]: I1205 10:38:14.142383 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-4vxng" Dec 05 10:38:14.158531 master-0 kubenswrapper[10444]: I1205 10:38:14.145904 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:38:14.158531 master-0 kubenswrapper[10444]: I1205 10:38:14.147380 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-2-master-0" event={"ID":"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35","Type":"ContainerStarted","Data":"4751d52074fab20e20ae168d59e19aac68268e7927466f42d617e9b690cf2dab"} Dec 05 10:38:14.183449 master-0 kubenswrapper[10444]: I1205 10:38:14.182485 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" event={"ID":"82ef99d4-41b7-4859-a5a1-4e31ce614a2a","Type":"ContainerDied","Data":"0f8b8ef2f443d099f39fe59fdd4477b3381b4efa500cd40773a08349e311ef2c"} Dec 05 10:38:14.183449 master-0 kubenswrapper[10444]: I1205 10:38:14.182558 10444 scope.go:117] "RemoveContainer" containerID="8575f477d29594718db712e5d56ee7bf4ba0a8c565ae4540b71d0eb7f5d45356" Dec 05 10:38:14.183449 master-0 kubenswrapper[10444]: I1205 10:38:14.182624 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m" Dec 05 10:38:14.205448 master-0 kubenswrapper[10444]: I1205 10:38:14.205110 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" event={"ID":"2118a526-c5ce-4c64-b955-48629a5db35b","Type":"ContainerStarted","Data":"126499044c53fa332cf0bb9d0a9b5af5164e53dfca2e288822486c026a37849e"} Dec 05 10:38:14.245392 master-0 kubenswrapper[10444]: I1205 10:38:14.245344 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" event={"ID":"941ce297-c739-41e4-804e-fd7d83726170","Type":"ContainerStarted","Data":"e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e"} Dec 05 10:38:14.246180 master-0 kubenswrapper[10444]: I1205 10:38:14.246135 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:14.256991 master-0 kubenswrapper[10444]: I1205 10:38:14.256915 10444 generic.go:334] "Generic (PLEG): container finished" podID="d5198438-06ae-4e63-a7e3-950ba23bba9c" containerID="4aeba45988671d42c201d0e79c4b96e55ead796c0fe34250d02d4c4bd0c61ad7" exitCode=0 Dec 05 10:38:14.257189 master-0 kubenswrapper[10444]: I1205 10:38:14.257010 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" event={"ID":"d5198438-06ae-4e63-a7e3-950ba23bba9c","Type":"ContainerDied","Data":"4aeba45988671d42c201d0e79c4b96e55ead796c0fe34250d02d4c4bd0c61ad7"} Dec 05 10:38:14.257610 master-0 kubenswrapper[10444]: I1205 10:38:14.257552 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-4vxng" podStartSLOduration=5.5558085 podStartE2EDuration="17.257531437s" podCreationTimestamp="2025-12-05 10:37:57 +0000 UTC" firstStartedPulling="2025-12-05 10:38:01.715494543 +0000 UTC m=+22.307407110" lastFinishedPulling="2025-12-05 10:38:13.41721748 +0000 UTC m=+34.009130047" observedRunningTime="2025-12-05 10:38:14.234883737 +0000 UTC m=+34.826796304" watchObservedRunningTime="2025-12-05 10:38:14.257531437 +0000 UTC m=+34.849444004" Dec 05 10:38:14.258267 master-0 kubenswrapper[10444]: I1205 10:38:14.258243 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m"] Dec 05 10:38:14.263549 master-0 kubenswrapper[10444]: I1205 10:38:14.262402 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-cluster-version/cluster-version-operator-77dfcc565f-bv84m"] Dec 05 10:38:14.263549 master-0 kubenswrapper[10444]: I1205 10:38:14.262727 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:14.263549 master-0 kubenswrapper[10444]: I1205 10:38:14.263117 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" event={"ID":"a4fd453c-e667-4bdc-aa9e-3d95ff707200","Type":"ContainerStarted","Data":"80b1488f509c166b5b116f44dccb9102304f263ee75bb4a43991cef27c102c84"} Dec 05 10:38:14.279728 master-0 kubenswrapper[10444]: I1205 10:38:14.274084 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" event={"ID":"592373ae-a952-4dd3-a9bc-f9c9c19c0802","Type":"ContainerStarted","Data":"2c8d17056f01fe2b63601e7d3f2b689f6a1b5945d124ed1dd3ebcdd26090636d"} Dec 05 10:38:14.279728 master-0 kubenswrapper[10444]: I1205 10:38:14.276381 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" event={"ID":"ca45c52e-fb30-4e7c-8c3f-e685c0909916","Type":"ContainerStarted","Data":"bd5095616b6a8494a79971c731d5ef49c3bae5cbd4111190f1e5dfde0e3d9554"} Dec 05 10:38:14.279728 master-0 kubenswrapper[10444]: I1205 10:38:14.279662 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" event={"ID":"8c649a16-c187-412e-b5da-62a00bee38ab","Type":"ContainerStarted","Data":"f52680e32d7f41ad89c986da65bb1e4f54d36a47dc2f03c6fe8b9d436655d9be"} Dec 05 10:38:14.288894 master-0 kubenswrapper[10444]: I1205 10:38:14.280327 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:38:14.316339 master-0 kubenswrapper[10444]: I1205 10:38:14.316269 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" podStartSLOduration=3.860800263 podStartE2EDuration="15.316244887s" podCreationTimestamp="2025-12-05 10:37:59 +0000 UTC" firstStartedPulling="2025-12-05 10:38:01.963235837 +0000 UTC m=+22.555148404" lastFinishedPulling="2025-12-05 10:38:13.418680461 +0000 UTC m=+34.010593028" observedRunningTime="2025-12-05 10:38:14.294019088 +0000 UTC m=+34.885931655" watchObservedRunningTime="2025-12-05 10:38:14.316244887 +0000 UTC m=+34.908157454" Dec 05 10:38:14.317072 master-0 kubenswrapper[10444]: I1205 10:38:14.317022 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6"] Dec 05 10:38:14.317260 master-0 kubenswrapper[10444]: E1205 10:38:14.317229 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ef99d4-41b7-4859-a5a1-4e31ce614a2a" containerName="cluster-version-operator" Dec 05 10:38:14.317260 master-0 kubenswrapper[10444]: I1205 10:38:14.317247 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ef99d4-41b7-4859-a5a1-4e31ce614a2a" containerName="cluster-version-operator" Dec 05 10:38:14.317351 master-0 kubenswrapper[10444]: I1205 10:38:14.317341 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ef99d4-41b7-4859-a5a1-4e31ce614a2a" containerName="cluster-version-operator" Dec 05 10:38:14.317737 master-0 kubenswrapper[10444]: I1205 10:38:14.317705 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.323957 master-0 kubenswrapper[10444]: I1205 10:38:14.323899 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 10:38:14.324143 master-0 kubenswrapper[10444]: I1205 10:38:14.324119 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 10:38:14.330531 master-0 kubenswrapper[10444]: I1205 10:38:14.330467 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 10:38:14.406464 master-0 kubenswrapper[10444]: I1205 10:38:14.406372 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fd0a349-0f66-4c85-95e4-8bbc96648c05-serving-cert\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.406464 master-0 kubenswrapper[10444]: I1205 10:38:14.406476 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-cvo-updatepayloads\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.406875 master-0 kubenswrapper[10444]: I1205 10:38:14.406518 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-ssl-certs\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.406875 master-0 kubenswrapper[10444]: I1205 10:38:14.406545 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1fd0a349-0f66-4c85-95e4-8bbc96648c05-kube-api-access\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.406875 master-0 kubenswrapper[10444]: I1205 10:38:14.406571 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1fd0a349-0f66-4c85-95e4-8bbc96648c05-service-ca\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.509280 master-0 kubenswrapper[10444]: I1205 10:38:14.509196 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-cvo-updatepayloads\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.509749 master-0 kubenswrapper[10444]: I1205 10:38:14.509380 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-cvo-updatepayloads\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.509954 master-0 kubenswrapper[10444]: I1205 10:38:14.509851 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-ssl-certs\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.509954 master-0 kubenswrapper[10444]: I1205 10:38:14.509890 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1fd0a349-0f66-4c85-95e4-8bbc96648c05-kube-api-access\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.509954 master-0 kubenswrapper[10444]: I1205 10:38:14.509914 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1fd0a349-0f66-4c85-95e4-8bbc96648c05-service-ca\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.510055 master-0 kubenswrapper[10444]: I1205 10:38:14.509997 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-ssl-certs\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.510335 master-0 kubenswrapper[10444]: I1205 10:38:14.510118 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fd0a349-0f66-4c85-95e4-8bbc96648c05-serving-cert\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.511288 master-0 kubenswrapper[10444]: I1205 10:38:14.511247 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1fd0a349-0f66-4c85-95e4-8bbc96648c05-service-ca\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.516370 master-0 kubenswrapper[10444]: I1205 10:38:14.516295 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fd0a349-0f66-4c85-95e4-8bbc96648c05-serving-cert\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.531146 master-0 kubenswrapper[10444]: I1205 10:38:14.530991 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1fd0a349-0f66-4c85-95e4-8bbc96648c05-kube-api-access\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:14.696817 master-0 kubenswrapper[10444]: I1205 10:38:14.696770 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:38:15.291087 master-0 kubenswrapper[10444]: I1205 10:38:15.291030 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-2-master-0" event={"ID":"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35","Type":"ContainerStarted","Data":"e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c"} Dec 05 10:38:15.294199 master-0 kubenswrapper[10444]: I1205 10:38:15.294152 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" event={"ID":"1fd0a349-0f66-4c85-95e4-8bbc96648c05","Type":"ContainerStarted","Data":"465abd54a7076d7b78aa23778511467ea5203eaf00e16f71acec306b149b36f4"} Dec 05 10:38:15.294199 master-0 kubenswrapper[10444]: I1205 10:38:15.294195 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" event={"ID":"1fd0a349-0f66-4c85-95e4-8bbc96648c05","Type":"ContainerStarted","Data":"0e8ac47ff99876f6c126e479233a75b5de6a9aabecbf3b414116156cf0e23f86"} Dec 05 10:38:15.297068 master-0 kubenswrapper[10444]: I1205 10:38:15.297025 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" event={"ID":"d5198438-06ae-4e63-a7e3-950ba23bba9c","Type":"ContainerStarted","Data":"4ae666475263cd4c6567beb58967d8630a346a399af0e4cabc03f8e1a2cf27d1"} Dec 05 10:38:15.297144 master-0 kubenswrapper[10444]: I1205 10:38:15.297094 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" event={"ID":"d5198438-06ae-4e63-a7e3-950ba23bba9c","Type":"ContainerStarted","Data":"76cd1d33ba4a7969cc5a2ff41bc2183002513f068f07fe097d458ccd9787e47e"} Dec 05 10:38:15.299945 master-0 kubenswrapper[10444]: I1205 10:38:15.299903 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" event={"ID":"a4fd453c-e667-4bdc-aa9e-3d95ff707200","Type":"ContainerStarted","Data":"2e9155fd74fd1d13dcef2794db107599bb6fa96cc533e3303f4bbbfdbd8c0208"} Dec 05 10:38:15.300022 master-0 kubenswrapper[10444]: I1205 10:38:15.299947 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" event={"ID":"a4fd453c-e667-4bdc-aa9e-3d95ff707200","Type":"ContainerStarted","Data":"08630bfb9ddeb1f85d444e0d89244867195f7b2cfb10ecc8ec97020808d3e73c"} Dec 05 10:38:15.300101 master-0 kubenswrapper[10444]: I1205 10:38:15.300070 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:15.301677 master-0 kubenswrapper[10444]: I1205 10:38:15.301649 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" event={"ID":"592373ae-a952-4dd3-a9bc-f9c9c19c0802","Type":"ContainerStarted","Data":"42ab2336c6be11ec1661443ac87e4b3c0d5e6ea08c7b98227a204a1767ffdd5a"} Dec 05 10:38:15.301769 master-0 kubenswrapper[10444]: I1205 10:38:15.301702 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" event={"ID":"592373ae-a952-4dd3-a9bc-f9c9c19c0802","Type":"ContainerStarted","Data":"665b41c4b05327dcf1d9daca6aa21544857362b43e6c7d8ef7fac95d7153f9b2"} Dec 05 10:38:15.318064 master-0 kubenswrapper[10444]: I1205 10:38:15.317983 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/installer-2-master-0" podStartSLOduration=5.317955957 podStartE2EDuration="5.317955957s" podCreationTimestamp="2025-12-05 10:38:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:15.308244733 +0000 UTC m=+35.900157310" watchObservedRunningTime="2025-12-05 10:38:15.317955957 +0000 UTC m=+35.909868524" Dec 05 10:38:15.352018 master-0 kubenswrapper[10444]: I1205 10:38:15.351929 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" podStartSLOduration=10.573054209 podStartE2EDuration="22.351897907s" podCreationTimestamp="2025-12-05 10:37:53 +0000 UTC" firstStartedPulling="2025-12-05 10:38:01.673111864 +0000 UTC m=+22.265024431" lastFinishedPulling="2025-12-05 10:38:13.451955562 +0000 UTC m=+34.043868129" observedRunningTime="2025-12-05 10:38:15.336675276 +0000 UTC m=+35.928587843" watchObservedRunningTime="2025-12-05 10:38:15.351897907 +0000 UTC m=+35.943810474" Dec 05 10:38:15.386455 master-0 kubenswrapper[10444]: I1205 10:38:15.384698 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" podStartSLOduration=1.3846615230000001 podStartE2EDuration="1.384661523s" podCreationTimestamp="2025-12-05 10:38:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:15.353371328 +0000 UTC m=+35.945283915" watchObservedRunningTime="2025-12-05 10:38:15.384661523 +0000 UTC m=+35.976574090" Dec 05 10:38:15.417573 master-0 kubenswrapper[10444]: I1205 10:38:15.417485 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" podStartSLOduration=7.417465261 podStartE2EDuration="7.417465261s" podCreationTimestamp="2025-12-05 10:38:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:15.386826424 +0000 UTC m=+35.978738991" watchObservedRunningTime="2025-12-05 10:38:15.417465261 +0000 UTC m=+36.009377838" Dec 05 10:38:15.604450 master-0 kubenswrapper[10444]: I1205 10:38:15.604282 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82ef99d4-41b7-4859-a5a1-4e31ce614a2a" path="/var/lib/kubelet/pods/82ef99d4-41b7-4859-a5a1-4e31ce614a2a/volumes" Dec 05 10:38:16.308633 master-0 kubenswrapper[10444]: I1205 10:38:16.308575 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:16.556013 master-0 kubenswrapper[10444]: I1205 10:38:16.555907 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" podStartSLOduration=8.555888816 podStartE2EDuration="8.555888816s" podCreationTimestamp="2025-12-05 10:38:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:15.417134361 +0000 UTC m=+36.009046928" watchObservedRunningTime="2025-12-05 10:38:16.555888816 +0000 UTC m=+37.147801393" Dec 05 10:38:16.558961 master-0 kubenswrapper[10444]: I1205 10:38:16.558866 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-scheduler/installer-2-master-0"] Dec 05 10:38:17.314140 master-0 kubenswrapper[10444]: I1205 10:38:17.314091 10444 generic.go:334] "Generic (PLEG): container finished" podID="ca45c52e-fb30-4e7c-8c3f-e685c0909916" containerID="0bd0566a23e122fff36d9dfda48e7737f21d3d5b04a9764846004be6154bdb8d" exitCode=0 Dec 05 10:38:17.314751 master-0 kubenswrapper[10444]: I1205 10:38:17.314170 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" event={"ID":"ca45c52e-fb30-4e7c-8c3f-e685c0909916","Type":"ContainerDied","Data":"0bd0566a23e122fff36d9dfda48e7737f21d3d5b04a9764846004be6154bdb8d"} Dec 05 10:38:17.318348 master-0 kubenswrapper[10444]: I1205 10:38:17.318299 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-scheduler/installer-2-master-0" podUID="9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35" containerName="installer" containerID="cri-o://e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c" gracePeriod=30 Dec 05 10:38:17.318612 master-0 kubenswrapper[10444]: I1205 10:38:17.318577 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" event={"ID":"2118a526-c5ce-4c64-b955-48629a5db35b","Type":"ContainerStarted","Data":"4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675"} Dec 05 10:38:17.319183 master-0 kubenswrapper[10444]: I1205 10:38:17.319104 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:17.370094 master-0 kubenswrapper[10444]: I1205 10:38:17.368612 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" podStartSLOduration=15.554579595 podStartE2EDuration="18.368595202s" podCreationTimestamp="2025-12-05 10:37:59 +0000 UTC" firstStartedPulling="2025-12-05 10:38:14.124602499 +0000 UTC m=+34.716515066" lastFinishedPulling="2025-12-05 10:38:16.938618106 +0000 UTC m=+37.530530673" observedRunningTime="2025-12-05 10:38:17.368585892 +0000 UTC m=+37.960498469" watchObservedRunningTime="2025-12-05 10:38:17.368595202 +0000 UTC m=+37.960507769" Dec 05 10:38:17.705362 master-0 kubenswrapper[10444]: I1205 10:38:17.705112 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:17.977006 master-0 kubenswrapper[10444]: I1205 10:38:17.976758 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler_installer-2-master-0_9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35/installer/0.log" Dec 05 10:38:17.977006 master-0 kubenswrapper[10444]: I1205 10:38:17.976819 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:18.076146 master-0 kubenswrapper[10444]: I1205 10:38:18.076020 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-var-lock\") pod \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " Dec 05 10:38:18.076347 master-0 kubenswrapper[10444]: I1205 10:38:18.076319 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-var-lock" (OuterVolumeSpecName: "var-lock") pod "9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35" (UID: "9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:38:18.176988 master-0 kubenswrapper[10444]: I1205 10:38:18.176715 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kube-api-access\") pod \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " Dec 05 10:38:18.176988 master-0 kubenswrapper[10444]: I1205 10:38:18.176782 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kubelet-dir\") pod \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\" (UID: \"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35\") " Dec 05 10:38:18.176988 master-0 kubenswrapper[10444]: I1205 10:38:18.176902 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:18.176988 master-0 kubenswrapper[10444]: I1205 10:38:18.176949 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35" (UID: "9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:38:18.180806 master-0 kubenswrapper[10444]: I1205 10:38:18.180765 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35" (UID: "9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:38:18.277721 master-0 kubenswrapper[10444]: I1205 10:38:18.277653 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:18.277721 master-0 kubenswrapper[10444]: I1205 10:38:18.277716 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:18.325246 master-0 kubenswrapper[10444]: I1205 10:38:18.324028 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler_installer-2-master-0_9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35/installer/0.log" Dec 05 10:38:18.325246 master-0 kubenswrapper[10444]: I1205 10:38:18.324078 10444 generic.go:334] "Generic (PLEG): container finished" podID="9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35" containerID="e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c" exitCode=1 Dec 05 10:38:18.325246 master-0 kubenswrapper[10444]: I1205 10:38:18.324175 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-2-master-0" Dec 05 10:38:18.325246 master-0 kubenswrapper[10444]: I1205 10:38:18.324642 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-2-master-0" event={"ID":"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35","Type":"ContainerDied","Data":"e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c"} Dec 05 10:38:18.325246 master-0 kubenswrapper[10444]: I1205 10:38:18.324708 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-2-master-0" event={"ID":"9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35","Type":"ContainerDied","Data":"4751d52074fab20e20ae168d59e19aac68268e7927466f42d617e9b690cf2dab"} Dec 05 10:38:18.325246 master-0 kubenswrapper[10444]: I1205 10:38:18.324731 10444 scope.go:117] "RemoveContainer" containerID="e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c" Dec 05 10:38:18.327311 master-0 kubenswrapper[10444]: I1205 10:38:18.327265 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" event={"ID":"ca45c52e-fb30-4e7c-8c3f-e685c0909916","Type":"ContainerStarted","Data":"c032d202c89f9b16c0498b0a51e62c4cede2bf746a86bd5cb7131a6006bf4679"} Dec 05 10:38:18.340416 master-0 kubenswrapper[10444]: I1205 10:38:18.340388 10444 scope.go:117] "RemoveContainer" containerID="e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c" Dec 05 10:38:18.342390 master-0 kubenswrapper[10444]: E1205 10:38:18.342329 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c\": container with ID starting with e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c not found: ID does not exist" containerID="e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c" Dec 05 10:38:18.342511 master-0 kubenswrapper[10444]: I1205 10:38:18.342386 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c"} err="failed to get container status \"e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c\": rpc error: code = NotFound desc = could not find container \"e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c\": container with ID starting with e2442b493963516a3ec33530cdbe5e21d88e812ea930bbbbd413bc1bc5f2526c not found: ID does not exist" Dec 05 10:38:18.790793 master-0 kubenswrapper[10444]: I1205 10:38:18.790701 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/installer-3-master-0"] Dec 05 10:38:18.791105 master-0 kubenswrapper[10444]: E1205 10:38:18.790944 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35" containerName="installer" Dec 05 10:38:18.791105 master-0 kubenswrapper[10444]: I1205 10:38:18.790970 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35" containerName="installer" Dec 05 10:38:18.791170 master-0 kubenswrapper[10444]: I1205 10:38:18.791107 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35" containerName="installer" Dec 05 10:38:18.791567 master-0 kubenswrapper[10444]: I1205 10:38:18.791546 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:38:18.823688 master-0 kubenswrapper[10444]: I1205 10:38:18.823631 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:38:18.823688 master-0 kubenswrapper[10444]: I1205 10:38:18.823687 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:38:18.984922 master-0 kubenswrapper[10444]: I1205 10:38:18.984852 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:38:18.984922 master-0 kubenswrapper[10444]: I1205 10:38:18.984921 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-var-lock\") pod \"installer-3-master-0\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:38:18.985168 master-0 kubenswrapper[10444]: I1205 10:38:18.985066 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6325f9a2-86d5-4fac-9e71-331b2628f965-kube-api-access\") pod \"installer-3-master-0\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:38:19.085639 master-0 kubenswrapper[10444]: I1205 10:38:19.085517 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:38:19.085639 master-0 kubenswrapper[10444]: I1205 10:38:19.085577 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-var-lock\") pod \"installer-3-master-0\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:38:19.085639 master-0 kubenswrapper[10444]: I1205 10:38:19.085605 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6325f9a2-86d5-4fac-9e71-331b2628f965-kube-api-access\") pod \"installer-3-master-0\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:38:19.085883 master-0 kubenswrapper[10444]: I1205 10:38:19.085735 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:38:19.085945 master-0 kubenswrapper[10444]: I1205 10:38:19.085921 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-var-lock\") pod \"installer-3-master-0\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:38:19.585600 master-0 kubenswrapper[10444]: I1205 10:38:19.585515 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:38:19.891708 master-0 kubenswrapper[10444]: I1205 10:38:19.891584 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/installer-3-master-0"] Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: I1205 10:38:20.054309 10444 patch_prober.go:28] interesting pod/apiserver-5b9fd577f8-6sxcx container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]log ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]etcd ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]poststarthook/generic-apiserver-start-informers ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]poststarthook/max-in-flight-filter ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]poststarthook/project.openshift.io-projectcache ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]poststarthook/openshift.io-startinformers ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 05 10:38:20.054387 master-0 kubenswrapper[10444]: livez check failed Dec 05 10:38:20.055190 master-0 kubenswrapper[10444]: I1205 10:38:20.054436 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" podUID="d5198438-06ae-4e63-a7e3-950ba23bba9c" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:38:20.112449 master-0 kubenswrapper[10444]: I1205 10:38:20.111993 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6325f9a2-86d5-4fac-9e71-331b2628f965-kube-api-access\") pod \"installer-3-master-0\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:38:20.161514 master-0 kubenswrapper[10444]: I1205 10:38:20.160207 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-1-master-0"] Dec 05 10:38:20.161514 master-0 kubenswrapper[10444]: I1205 10:38:20.160999 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:38:20.167436 master-0 kubenswrapper[10444]: I1205 10:38:20.166056 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 10:38:20.207451 master-0 kubenswrapper[10444]: I1205 10:38:20.197414 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-1-master-0"] Dec 05 10:38:20.217601 master-0 kubenswrapper[10444]: I1205 10:38:20.210950 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kube-api-access\") pod \"installer-1-master-0\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:38:20.217601 master-0 kubenswrapper[10444]: I1205 10:38:20.211015 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:38:20.217601 master-0 kubenswrapper[10444]: I1205 10:38:20.211065 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-var-lock\") pod \"installer-1-master-0\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:38:20.226910 master-0 kubenswrapper[10444]: I1205 10:38:20.226851 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-scheduler/installer-2-master-0"] Dec 05 10:38:20.227793 master-0 kubenswrapper[10444]: I1205 10:38:20.227732 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-scheduler/installer-2-master-0"] Dec 05 10:38:20.263319 master-0 kubenswrapper[10444]: I1205 10:38:20.261687 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" podStartSLOduration=15.072314424 podStartE2EDuration="18.261664433s" podCreationTimestamp="2025-12-05 10:38:02 +0000 UTC" firstStartedPulling="2025-12-05 10:38:13.743062282 +0000 UTC m=+34.334974849" lastFinishedPulling="2025-12-05 10:38:16.932412291 +0000 UTC m=+37.524324858" observedRunningTime="2025-12-05 10:38:20.260869532 +0000 UTC m=+40.852782099" watchObservedRunningTime="2025-12-05 10:38:20.261664433 +0000 UTC m=+40.853577010" Dec 05 10:38:20.311187 master-0 kubenswrapper[10444]: I1205 10:38:20.311130 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:38:20.316714 master-0 kubenswrapper[10444]: I1205 10:38:20.314440 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kube-api-access\") pod \"installer-1-master-0\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:38:20.316714 master-0 kubenswrapper[10444]: I1205 10:38:20.314517 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:38:20.316714 master-0 kubenswrapper[10444]: I1205 10:38:20.314581 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-var-lock\") pod \"installer-1-master-0\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:38:20.316714 master-0 kubenswrapper[10444]: I1205 10:38:20.314669 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-var-lock\") pod \"installer-1-master-0\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:38:20.316714 master-0 kubenswrapper[10444]: I1205 10:38:20.315022 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:38:20.334332 master-0 kubenswrapper[10444]: I1205 10:38:20.325719 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6458c74b4c-4gvlc"] Dec 05 10:38:20.334332 master-0 kubenswrapper[10444]: I1205 10:38:20.325955 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" podUID="941ce297-c739-41e4-804e-fd7d83726170" containerName="controller-manager" containerID="cri-o://e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e" gracePeriod=30 Dec 05 10:38:20.337507 master-0 kubenswrapper[10444]: I1205 10:38:20.335184 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kube-api-access\") pod \"installer-1-master-0\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:38:20.352514 master-0 kubenswrapper[10444]: I1205 10:38:20.348802 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln"] Dec 05 10:38:20.352514 master-0 kubenswrapper[10444]: I1205 10:38:20.349456 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" podUID="2118a526-c5ce-4c64-b955-48629a5db35b" containerName="route-controller-manager" containerID="cri-o://4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675" gracePeriod=30 Dec 05 10:38:20.506961 master-0 kubenswrapper[10444]: I1205 10:38:20.506901 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:38:20.758773 master-0 kubenswrapper[10444]: I1205 10:38:20.758729 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:20.764634 master-0 kubenswrapper[10444]: I1205 10:38:20.764437 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:20.765452 master-0 kubenswrapper[10444]: I1205 10:38:20.765377 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/installer-3-master-0"] Dec 05 10:38:20.828598 master-0 kubenswrapper[10444]: I1205 10:38:20.828550 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-client-ca\") pod \"941ce297-c739-41e4-804e-fd7d83726170\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " Dec 05 10:38:20.828598 master-0 kubenswrapper[10444]: I1205 10:38:20.828595 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-config\") pod \"2118a526-c5ce-4c64-b955-48629a5db35b\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " Dec 05 10:38:20.828800 master-0 kubenswrapper[10444]: I1205 10:38:20.828623 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/941ce297-c739-41e4-804e-fd7d83726170-serving-cert\") pod \"941ce297-c739-41e4-804e-fd7d83726170\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " Dec 05 10:38:20.828800 master-0 kubenswrapper[10444]: I1205 10:38:20.828644 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnmjt\" (UniqueName: \"kubernetes.io/projected/941ce297-c739-41e4-804e-fd7d83726170-kube-api-access-bnmjt\") pod \"941ce297-c739-41e4-804e-fd7d83726170\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " Dec 05 10:38:20.828800 master-0 kubenswrapper[10444]: I1205 10:38:20.828697 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbl76\" (UniqueName: \"kubernetes.io/projected/2118a526-c5ce-4c64-b955-48629a5db35b-kube-api-access-rbl76\") pod \"2118a526-c5ce-4c64-b955-48629a5db35b\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " Dec 05 10:38:20.828800 master-0 kubenswrapper[10444]: I1205 10:38:20.828726 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-config\") pod \"941ce297-c739-41e4-804e-fd7d83726170\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " Dec 05 10:38:20.828950 master-0 kubenswrapper[10444]: I1205 10:38:20.828913 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2118a526-c5ce-4c64-b955-48629a5db35b-serving-cert\") pod \"2118a526-c5ce-4c64-b955-48629a5db35b\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " Dec 05 10:38:20.828992 master-0 kubenswrapper[10444]: I1205 10:38:20.828974 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-proxy-ca-bundles\") pod \"941ce297-c739-41e4-804e-fd7d83726170\" (UID: \"941ce297-c739-41e4-804e-fd7d83726170\") " Dec 05 10:38:20.829052 master-0 kubenswrapper[10444]: I1205 10:38:20.829030 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-client-ca\") pod \"2118a526-c5ce-4c64-b955-48629a5db35b\" (UID: \"2118a526-c5ce-4c64-b955-48629a5db35b\") " Dec 05 10:38:20.829184 master-0 kubenswrapper[10444]: I1205 10:38:20.829142 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-config" (OuterVolumeSpecName: "config") pod "2118a526-c5ce-4c64-b955-48629a5db35b" (UID: "2118a526-c5ce-4c64-b955-48629a5db35b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:38:20.829411 master-0 kubenswrapper[10444]: I1205 10:38:20.829389 10444 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:20.829902 master-0 kubenswrapper[10444]: I1205 10:38:20.829873 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-config" (OuterVolumeSpecName: "config") pod "941ce297-c739-41e4-804e-fd7d83726170" (UID: "941ce297-c739-41e4-804e-fd7d83726170"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:38:20.829948 master-0 kubenswrapper[10444]: I1205 10:38:20.829913 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-client-ca" (OuterVolumeSpecName: "client-ca") pod "941ce297-c739-41e4-804e-fd7d83726170" (UID: "941ce297-c739-41e4-804e-fd7d83726170"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:38:20.830037 master-0 kubenswrapper[10444]: I1205 10:38:20.829981 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "941ce297-c739-41e4-804e-fd7d83726170" (UID: "941ce297-c739-41e4-804e-fd7d83726170"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:38:20.830085 master-0 kubenswrapper[10444]: I1205 10:38:20.829980 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-client-ca" (OuterVolumeSpecName: "client-ca") pod "2118a526-c5ce-4c64-b955-48629a5db35b" (UID: "2118a526-c5ce-4c64-b955-48629a5db35b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:38:20.832237 master-0 kubenswrapper[10444]: I1205 10:38:20.832193 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/941ce297-c739-41e4-804e-fd7d83726170-kube-api-access-bnmjt" (OuterVolumeSpecName: "kube-api-access-bnmjt") pod "941ce297-c739-41e4-804e-fd7d83726170" (UID: "941ce297-c739-41e4-804e-fd7d83726170"). InnerVolumeSpecName "kube-api-access-bnmjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:38:20.832358 master-0 kubenswrapper[10444]: I1205 10:38:20.832329 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2118a526-c5ce-4c64-b955-48629a5db35b-kube-api-access-rbl76" (OuterVolumeSpecName: "kube-api-access-rbl76") pod "2118a526-c5ce-4c64-b955-48629a5db35b" (UID: "2118a526-c5ce-4c64-b955-48629a5db35b"). InnerVolumeSpecName "kube-api-access-rbl76". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:38:20.832487 master-0 kubenswrapper[10444]: I1205 10:38:20.832448 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2118a526-c5ce-4c64-b955-48629a5db35b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2118a526-c5ce-4c64-b955-48629a5db35b" (UID: "2118a526-c5ce-4c64-b955-48629a5db35b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:38:20.832705 master-0 kubenswrapper[10444]: I1205 10:38:20.832676 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/941ce297-c739-41e4-804e-fd7d83726170-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "941ce297-c739-41e4-804e-fd7d83726170" (UID: "941ce297-c739-41e4-804e-fd7d83726170"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:38:20.930315 master-0 kubenswrapper[10444]: I1205 10:38:20.930254 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnmjt\" (UniqueName: \"kubernetes.io/projected/941ce297-c739-41e4-804e-fd7d83726170-kube-api-access-bnmjt\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:20.930315 master-0 kubenswrapper[10444]: I1205 10:38:20.930297 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbl76\" (UniqueName: \"kubernetes.io/projected/2118a526-c5ce-4c64-b955-48629a5db35b-kube-api-access-rbl76\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:20.930315 master-0 kubenswrapper[10444]: I1205 10:38:20.930309 10444 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:20.930315 master-0 kubenswrapper[10444]: I1205 10:38:20.930320 10444 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2118a526-c5ce-4c64-b955-48629a5db35b-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:20.930315 master-0 kubenswrapper[10444]: I1205 10:38:20.930329 10444 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-proxy-ca-bundles\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:20.930315 master-0 kubenswrapper[10444]: I1205 10:38:20.930337 10444 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2118a526-c5ce-4c64-b955-48629a5db35b-client-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:20.930748 master-0 kubenswrapper[10444]: I1205 10:38:20.930346 10444 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/941ce297-c739-41e4-804e-fd7d83726170-client-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:20.930748 master-0 kubenswrapper[10444]: I1205 10:38:20.930356 10444 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/941ce297-c739-41e4-804e-fd7d83726170-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:20.941760 master-0 kubenswrapper[10444]: I1205 10:38:20.939855 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-1-master-0"] Dec 05 10:38:20.948959 master-0 kubenswrapper[10444]: W1205 10:38:20.948910 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod5d8c854f_31f8_42d3_b28c_82c7d1b14ee9.slice/crio-0127ee22dc31281cc8a51e81727d0320a63ff8209b1fadb0e73c9f87ae0d74aa WatchSource:0}: Error finding container 0127ee22dc31281cc8a51e81727d0320a63ff8209b1fadb0e73c9f87ae0d74aa: Status 404 returned error can't find the container with id 0127ee22dc31281cc8a51e81727d0320a63ff8209b1fadb0e73c9f87ae0d74aa Dec 05 10:38:21.013457 master-0 kubenswrapper[10444]: I1205 10:38:21.013305 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:21.013457 master-0 kubenswrapper[10444]: I1205 10:38:21.013378 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:21.020976 master-0 kubenswrapper[10444]: I1205 10:38:21.020937 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:21.359206 master-0 kubenswrapper[10444]: I1205 10:38:21.358279 10444 generic.go:334] "Generic (PLEG): container finished" podID="941ce297-c739-41e4-804e-fd7d83726170" containerID="e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e" exitCode=0 Dec 05 10:38:21.359206 master-0 kubenswrapper[10444]: I1205 10:38:21.358341 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" event={"ID":"941ce297-c739-41e4-804e-fd7d83726170","Type":"ContainerDied","Data":"e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e"} Dec 05 10:38:21.359206 master-0 kubenswrapper[10444]: I1205 10:38:21.358350 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" Dec 05 10:38:21.359206 master-0 kubenswrapper[10444]: I1205 10:38:21.358380 10444 scope.go:117] "RemoveContainer" containerID="e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e" Dec 05 10:38:21.359206 master-0 kubenswrapper[10444]: I1205 10:38:21.358368 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6458c74b4c-4gvlc" event={"ID":"941ce297-c739-41e4-804e-fd7d83726170","Type":"ContainerDied","Data":"1e491ceac32476146ec752c9cf678575f2f44cfeb6d755985ac117ae34fbb8ac"} Dec 05 10:38:21.360703 master-0 kubenswrapper[10444]: I1205 10:38:21.360352 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-3-master-0" event={"ID":"6325f9a2-86d5-4fac-9e71-331b2628f965","Type":"ContainerStarted","Data":"df5bdcc899d9a7c7268a300825923259315bb7a02282ed76b8b7644641b06995"} Dec 05 10:38:21.360703 master-0 kubenswrapper[10444]: I1205 10:38:21.360387 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-3-master-0" event={"ID":"6325f9a2-86d5-4fac-9e71-331b2628f965","Type":"ContainerStarted","Data":"ba7c2d250edf116aab8b9c6bc4e16427c460bc08a2f3f72ce479a134eaf54756"} Dec 05 10:38:21.362606 master-0 kubenswrapper[10444]: I1205 10:38:21.361850 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-1-master-0" event={"ID":"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9","Type":"ContainerStarted","Data":"0d4c12890814d7bd72e2d33d03a6131a3d52c0e2122d3701215dc96f26a231b2"} Dec 05 10:38:21.362606 master-0 kubenswrapper[10444]: I1205 10:38:21.361888 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-1-master-0" event={"ID":"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9","Type":"ContainerStarted","Data":"0127ee22dc31281cc8a51e81727d0320a63ff8209b1fadb0e73c9f87ae0d74aa"} Dec 05 10:38:21.364177 master-0 kubenswrapper[10444]: I1205 10:38:21.364126 10444 generic.go:334] "Generic (PLEG): container finished" podID="2118a526-c5ce-4c64-b955-48629a5db35b" containerID="4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675" exitCode=0 Dec 05 10:38:21.364177 master-0 kubenswrapper[10444]: I1205 10:38:21.364150 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" Dec 05 10:38:21.364298 master-0 kubenswrapper[10444]: I1205 10:38:21.364166 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" event={"ID":"2118a526-c5ce-4c64-b955-48629a5db35b","Type":"ContainerDied","Data":"4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675"} Dec 05 10:38:21.364342 master-0 kubenswrapper[10444]: I1205 10:38:21.364294 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln" event={"ID":"2118a526-c5ce-4c64-b955-48629a5db35b","Type":"ContainerDied","Data":"126499044c53fa332cf0bb9d0a9b5af5164e53dfca2e288822486c026a37849e"} Dec 05 10:38:21.368369 master-0 kubenswrapper[10444]: I1205 10:38:21.368329 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:38:21.370408 master-0 kubenswrapper[10444]: I1205 10:38:21.370368 10444 scope.go:117] "RemoveContainer" containerID="e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e" Dec 05 10:38:21.370794 master-0 kubenswrapper[10444]: E1205 10:38:21.370752 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e\": container with ID starting with e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e not found: ID does not exist" containerID="e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e" Dec 05 10:38:21.370837 master-0 kubenswrapper[10444]: I1205 10:38:21.370802 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e"} err="failed to get container status \"e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e\": rpc error: code = NotFound desc = could not find container \"e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e\": container with ID starting with e4b8e38b475096650c9862e6e7326dcda5595b1ecd58a1f238315ebba783f34e not found: ID does not exist" Dec 05 10:38:21.370872 master-0 kubenswrapper[10444]: I1205 10:38:21.370839 10444 scope.go:117] "RemoveContainer" containerID="4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675" Dec 05 10:38:21.383661 master-0 kubenswrapper[10444]: I1205 10:38:21.383574 10444 scope.go:117] "RemoveContainer" containerID="4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675" Dec 05 10:38:21.386176 master-0 kubenswrapper[10444]: I1205 10:38:21.385977 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/installer-3-master-0" podStartSLOduration=3.385963459 podStartE2EDuration="3.385963459s" podCreationTimestamp="2025-12-05 10:38:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:21.385495676 +0000 UTC m=+41.977408243" watchObservedRunningTime="2025-12-05 10:38:21.385963459 +0000 UTC m=+41.977876026" Dec 05 10:38:21.386940 master-0 kubenswrapper[10444]: E1205 10:38:21.386886 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675\": container with ID starting with 4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675 not found: ID does not exist" containerID="4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675" Dec 05 10:38:21.387028 master-0 kubenswrapper[10444]: I1205 10:38:21.386950 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675"} err="failed to get container status \"4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675\": rpc error: code = NotFound desc = could not find container \"4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675\": container with ID starting with 4aaad930297436daa6b7a98b90e8394bb4628723f7101469fa8977b360bf1675 not found: ID does not exist" Dec 05 10:38:21.398399 master-0 kubenswrapper[10444]: I1205 10:38:21.398341 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-86f4478dbf-jqlt9"] Dec 05 10:38:21.398601 master-0 kubenswrapper[10444]: E1205 10:38:21.398568 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2118a526-c5ce-4c64-b955-48629a5db35b" containerName="route-controller-manager" Dec 05 10:38:21.398601 master-0 kubenswrapper[10444]: I1205 10:38:21.398592 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="2118a526-c5ce-4c64-b955-48629a5db35b" containerName="route-controller-manager" Dec 05 10:38:21.398700 master-0 kubenswrapper[10444]: E1205 10:38:21.398609 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="941ce297-c739-41e4-804e-fd7d83726170" containerName="controller-manager" Dec 05 10:38:21.398700 master-0 kubenswrapper[10444]: I1205 10:38:21.398618 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="941ce297-c739-41e4-804e-fd7d83726170" containerName="controller-manager" Dec 05 10:38:21.398794 master-0 kubenswrapper[10444]: I1205 10:38:21.398734 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="2118a526-c5ce-4c64-b955-48629a5db35b" containerName="route-controller-manager" Dec 05 10:38:21.398794 master-0 kubenswrapper[10444]: I1205 10:38:21.398755 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="941ce297-c739-41e4-804e-fd7d83726170" containerName="controller-manager" Dec 05 10:38:21.400088 master-0 kubenswrapper[10444]: I1205 10:38:21.399120 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.401387 master-0 kubenswrapper[10444]: I1205 10:38:21.401338 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln"] Dec 05 10:38:21.403334 master-0 kubenswrapper[10444]: I1205 10:38:21.402354 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 10:38:21.403334 master-0 kubenswrapper[10444]: I1205 10:38:21.402705 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 10:38:21.404988 master-0 kubenswrapper[10444]: I1205 10:38:21.404903 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 10:38:21.405101 master-0 kubenswrapper[10444]: I1205 10:38:21.405051 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 10:38:21.405851 master-0 kubenswrapper[10444]: I1205 10:38:21.405805 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 10:38:21.406301 master-0 kubenswrapper[10444]: I1205 10:38:21.406216 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f6f96665d-4nkln"] Dec 05 10:38:21.410116 master-0 kubenswrapper[10444]: I1205 10:38:21.410046 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 10:38:21.410613 master-0 kubenswrapper[10444]: I1205 10:38:21.410560 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-86f4478dbf-jqlt9"] Dec 05 10:38:21.437488 master-0 kubenswrapper[10444]: I1205 10:38:21.437412 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-proxy-ca-bundles\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.437734 master-0 kubenswrapper[10444]: I1205 10:38:21.437673 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-serving-cert\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.437773 master-0 kubenswrapper[10444]: I1205 10:38:21.437759 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95jmp\" (UniqueName: \"kubernetes.io/projected/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-kube-api-access-95jmp\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.437864 master-0 kubenswrapper[10444]: I1205 10:38:21.437826 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-client-ca\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.437917 master-0 kubenswrapper[10444]: I1205 10:38:21.437897 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-config\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.520151 master-0 kubenswrapper[10444]: I1205 10:38:21.485474 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6458c74b4c-4gvlc"] Dec 05 10:38:21.520151 master-0 kubenswrapper[10444]: I1205 10:38:21.506993 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6458c74b4c-4gvlc"] Dec 05 10:38:21.539023 master-0 kubenswrapper[10444]: I1205 10:38:21.538965 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95jmp\" (UniqueName: \"kubernetes.io/projected/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-kube-api-access-95jmp\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.539202 master-0 kubenswrapper[10444]: I1205 10:38:21.539037 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-client-ca\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.539202 master-0 kubenswrapper[10444]: I1205 10:38:21.539066 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-config\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.539202 master-0 kubenswrapper[10444]: I1205 10:38:21.539106 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-proxy-ca-bundles\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.539202 master-0 kubenswrapper[10444]: I1205 10:38:21.539157 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-serving-cert\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.540533 master-0 kubenswrapper[10444]: I1205 10:38:21.540503 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-client-ca\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.563447 master-0 kubenswrapper[10444]: I1205 10:38:21.562178 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-serving-cert\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.563922 master-0 kubenswrapper[10444]: I1205 10:38:21.563862 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-1-master-0" podStartSLOduration=1.563849508 podStartE2EDuration="1.563849508s" podCreationTimestamp="2025-12-05 10:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:21.562760628 +0000 UTC m=+42.154673195" watchObservedRunningTime="2025-12-05 10:38:21.563849508 +0000 UTC m=+42.155762075" Dec 05 10:38:21.574481 master-0 kubenswrapper[10444]: I1205 10:38:21.566371 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-proxy-ca-bundles\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.574481 master-0 kubenswrapper[10444]: I1205 10:38:21.570140 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-config\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.592449 master-0 kubenswrapper[10444]: I1205 10:38:21.592399 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95jmp\" (UniqueName: \"kubernetes.io/projected/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-kube-api-access-95jmp\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:21.603528 master-0 kubenswrapper[10444]: I1205 10:38:21.602786 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2118a526-c5ce-4c64-b955-48629a5db35b" path="/var/lib/kubelet/pods/2118a526-c5ce-4c64-b955-48629a5db35b/volumes" Dec 05 10:38:21.603528 master-0 kubenswrapper[10444]: I1205 10:38:21.603349 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="941ce297-c739-41e4-804e-fd7d83726170" path="/var/lib/kubelet/pods/941ce297-c739-41e4-804e-fd7d83726170/volumes" Dec 05 10:38:21.603878 master-0 kubenswrapper[10444]: I1205 10:38:21.603853 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35" path="/var/lib/kubelet/pods/9e7e6d3f-d579-4a5a-a0ad-bb4c6aa73a35/volumes" Dec 05 10:38:21.738296 master-0 kubenswrapper[10444]: I1205 10:38:21.738164 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:22.149728 master-0 kubenswrapper[10444]: I1205 10:38:22.149669 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-86f4478dbf-jqlt9"] Dec 05 10:38:22.159239 master-0 kubenswrapper[10444]: W1205 10:38:22.159170 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0cbad64_72b9_4ad3_9a42_4183e93c9ba0.slice/crio-0f19c2361e4856d55a1f7cfcf2bbf2e6aae6be0b4b861e0e68ba7772ccbc2487 WatchSource:0}: Error finding container 0f19c2361e4856d55a1f7cfcf2bbf2e6aae6be0b4b861e0e68ba7772ccbc2487: Status 404 returned error can't find the container with id 0f19c2361e4856d55a1f7cfcf2bbf2e6aae6be0b4b861e0e68ba7772ccbc2487 Dec 05 10:38:22.372114 master-0 kubenswrapper[10444]: I1205 10:38:22.372008 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" event={"ID":"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0","Type":"ContainerStarted","Data":"8aa1ac52d0f6c8b159f0da42738bae856c9e11f506fdfed545be3d79a9c5ba1d"} Dec 05 10:38:22.372114 master-0 kubenswrapper[10444]: I1205 10:38:22.372102 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" event={"ID":"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0","Type":"ContainerStarted","Data":"0f19c2361e4856d55a1f7cfcf2bbf2e6aae6be0b4b861e0e68ba7772ccbc2487"} Dec 05 10:38:22.372353 master-0 kubenswrapper[10444]: I1205 10:38:22.372244 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:22.374234 master-0 kubenswrapper[10444]: I1205 10:38:22.374178 10444 patch_prober.go:28] interesting pod/controller-manager-86f4478dbf-jqlt9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.128.0.45:8443/healthz\": dial tcp 10.128.0.45:8443: connect: connection refused" start-of-body= Dec 05 10:38:22.374301 master-0 kubenswrapper[10444]: I1205 10:38:22.374259 10444 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" probeResult="failure" output="Get \"https://10.128.0.45:8443/healthz\": dial tcp 10.128.0.45:8443: connect: connection refused" Dec 05 10:38:22.394673 master-0 kubenswrapper[10444]: I1205 10:38:22.394588 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" podStartSLOduration=2.394563965 podStartE2EDuration="2.394563965s" podCreationTimestamp="2025-12-05 10:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:22.394291077 +0000 UTC m=+42.986203704" watchObservedRunningTime="2025-12-05 10:38:22.394563965 +0000 UTC m=+42.986476552" Dec 05 10:38:22.402869 master-0 kubenswrapper[10444]: I1205 10:38:22.402751 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s"] Dec 05 10:38:22.403283 master-0 kubenswrapper[10444]: I1205 10:38:22.403258 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.406805 master-0 kubenswrapper[10444]: I1205 10:38:22.406754 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 10:38:22.406953 master-0 kubenswrapper[10444]: I1205 10:38:22.406887 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 10:38:22.407016 master-0 kubenswrapper[10444]: I1205 10:38:22.406965 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 10:38:22.407988 master-0 kubenswrapper[10444]: I1205 10:38:22.407941 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 10:38:22.408057 master-0 kubenswrapper[10444]: I1205 10:38:22.408030 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 10:38:22.415314 master-0 kubenswrapper[10444]: I1205 10:38:22.415240 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s"] Dec 05 10:38:22.450293 master-0 kubenswrapper[10444]: I1205 10:38:22.449674 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4lbg\" (UniqueName: \"kubernetes.io/projected/88cccb5b-1ad1-4fab-b34c-90252794ee20-kube-api-access-f4lbg\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.450293 master-0 kubenswrapper[10444]: I1205 10:38:22.450296 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-config\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.450845 master-0 kubenswrapper[10444]: I1205 10:38:22.450798 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-client-ca\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.451389 master-0 kubenswrapper[10444]: I1205 10:38:22.451303 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88cccb5b-1ad1-4fab-b34c-90252794ee20-serving-cert\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.553828 master-0 kubenswrapper[10444]: I1205 10:38:22.552910 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88cccb5b-1ad1-4fab-b34c-90252794ee20-serving-cert\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.554107 master-0 kubenswrapper[10444]: I1205 10:38:22.553877 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4lbg\" (UniqueName: \"kubernetes.io/projected/88cccb5b-1ad1-4fab-b34c-90252794ee20-kube-api-access-f4lbg\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.554107 master-0 kubenswrapper[10444]: I1205 10:38:22.553913 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-config\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.554107 master-0 kubenswrapper[10444]: I1205 10:38:22.553939 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-client-ca\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.555598 master-0 kubenswrapper[10444]: I1205 10:38:22.555551 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-config\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.555717 master-0 kubenswrapper[10444]: I1205 10:38:22.555651 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-client-ca\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.556673 master-0 kubenswrapper[10444]: I1205 10:38:22.556615 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88cccb5b-1ad1-4fab-b34c-90252794ee20-serving-cert\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.579845 master-0 kubenswrapper[10444]: I1205 10:38:22.579781 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4lbg\" (UniqueName: \"kubernetes.io/projected/88cccb5b-1ad1-4fab-b34c-90252794ee20-kube-api-access-f4lbg\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.738499 master-0 kubenswrapper[10444]: I1205 10:38:22.738372 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:22.753062 master-0 kubenswrapper[10444]: I1205 10:38:22.753003 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/installer-1-master-0"] Dec 05 10:38:22.753593 master-0 kubenswrapper[10444]: I1205 10:38:22.753561 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:38:22.755664 master-0 kubenswrapper[10444]: I1205 10:38:22.755630 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 05 10:38:22.767867 master-0 kubenswrapper[10444]: I1205 10:38:22.767800 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/installer-1-master-0"] Dec 05 10:38:22.860654 master-0 kubenswrapper[10444]: I1205 10:38:22.859922 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:38:22.860654 master-0 kubenswrapper[10444]: I1205 10:38:22.859999 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-var-lock\") pod \"installer-1-master-0\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:38:22.860654 master-0 kubenswrapper[10444]: I1205 10:38:22.860027 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kube-api-access\") pod \"installer-1-master-0\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:38:22.961505 master-0 kubenswrapper[10444]: I1205 10:38:22.961416 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:38:22.961505 master-0 kubenswrapper[10444]: I1205 10:38:22.961498 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-var-lock\") pod \"installer-1-master-0\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:38:22.961756 master-0 kubenswrapper[10444]: I1205 10:38:22.961531 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kube-api-access\") pod \"installer-1-master-0\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:38:22.961756 master-0 kubenswrapper[10444]: I1205 10:38:22.961609 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kubelet-dir\") pod \"installer-1-master-0\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:38:22.961876 master-0 kubenswrapper[10444]: I1205 10:38:22.961765 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-var-lock\") pod \"installer-1-master-0\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:38:22.978663 master-0 kubenswrapper[10444]: I1205 10:38:22.978547 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kube-api-access\") pod \"installer-1-master-0\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:38:23.094030 master-0 kubenswrapper[10444]: I1205 10:38:23.093962 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:38:23.177911 master-0 kubenswrapper[10444]: I1205 10:38:23.177843 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s"] Dec 05 10:38:23.279289 master-0 kubenswrapper[10444]: W1205 10:38:23.278924 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88cccb5b_1ad1_4fab_b34c_90252794ee20.slice/crio-1bde7edcbb5c913afcdfed1e9b165ffe883b909863e89e771b54560c18c0ac2a WatchSource:0}: Error finding container 1bde7edcbb5c913afcdfed1e9b165ffe883b909863e89e771b54560c18c0ac2a: Status 404 returned error can't find the container with id 1bde7edcbb5c913afcdfed1e9b165ffe883b909863e89e771b54560c18c0ac2a Dec 05 10:38:23.383534 master-0 kubenswrapper[10444]: I1205 10:38:23.381622 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" event={"ID":"88cccb5b-1ad1-4fab-b34c-90252794ee20","Type":"ContainerStarted","Data":"1bde7edcbb5c913afcdfed1e9b165ffe883b909863e89e771b54560c18c0ac2a"} Dec 05 10:38:23.386967 master-0 kubenswrapper[10444]: I1205 10:38:23.385929 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:38:23.417493 master-0 kubenswrapper[10444]: I1205 10:38:23.417456 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/installer-1-master-0"] Dec 05 10:38:23.443195 master-0 kubenswrapper[10444]: W1205 10:38:23.437028 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod738d9fe7_2d08_4aa4_842b_9cbeddf120fb.slice/crio-481bec329814be919f1e3dd3cd8fba26d8d823339446a0f9b66a76525352b23f WatchSource:0}: Error finding container 481bec329814be919f1e3dd3cd8fba26d8d823339446a0f9b66a76525352b23f: Status 404 returned error can't find the container with id 481bec329814be919f1e3dd3cd8fba26d8d823339446a0f9b66a76525352b23f Dec 05 10:38:23.499630 master-0 kubenswrapper[10444]: I1205 10:38:23.499586 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-4vxng" Dec 05 10:38:23.827197 master-0 kubenswrapper[10444]: I1205 10:38:23.827161 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:38:23.837776 master-0 kubenswrapper[10444]: I1205 10:38:23.830875 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:38:24.385898 master-0 kubenswrapper[10444]: I1205 10:38:24.385847 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" event={"ID":"88cccb5b-1ad1-4fab-b34c-90252794ee20","Type":"ContainerStarted","Data":"dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5"} Dec 05 10:38:24.386918 master-0 kubenswrapper[10444]: I1205 10:38:24.386896 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:24.388650 master-0 kubenswrapper[10444]: I1205 10:38:24.388624 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-1-master-0" event={"ID":"738d9fe7-2d08-4aa4-842b-9cbeddf120fb","Type":"ContainerStarted","Data":"8a8e52c1bfb7c50d31dbfbb099e72a94f8cf7f5c9eac30d42a9e7bb6f30721bf"} Dec 05 10:38:24.388650 master-0 kubenswrapper[10444]: I1205 10:38:24.388648 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-1-master-0" event={"ID":"738d9fe7-2d08-4aa4-842b-9cbeddf120fb","Type":"ContainerStarted","Data":"481bec329814be919f1e3dd3cd8fba26d8d823339446a0f9b66a76525352b23f"} Dec 05 10:38:24.397383 master-0 kubenswrapper[10444]: I1205 10:38:24.397347 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:38:24.445223 master-0 kubenswrapper[10444]: I1205 10:38:24.445143 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" podStartSLOduration=4.445118967 podStartE2EDuration="4.445118967s" podCreationTimestamp="2025-12-05 10:38:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:24.443642065 +0000 UTC m=+45.035554652" watchObservedRunningTime="2025-12-05 10:38:24.445118967 +0000 UTC m=+45.037031534" Dec 05 10:38:24.508512 master-0 kubenswrapper[10444]: I1205 10:38:24.506169 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/installer-1-master-0" podStartSLOduration=2.506149792 podStartE2EDuration="2.506149792s" podCreationTimestamp="2025-12-05 10:38:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:24.505620788 +0000 UTC m=+45.097533375" watchObservedRunningTime="2025-12-05 10:38:24.506149792 +0000 UTC m=+45.098062379" Dec 05 10:38:28.454956 master-0 kubenswrapper[10444]: I1205 10:38:28.454897 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t"] Dec 05 10:38:28.455639 master-0 kubenswrapper[10444]: I1205 10:38:28.455593 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:38:28.457908 master-0 kubenswrapper[10444]: I1205 10:38:28.457864 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 10:38:28.458064 master-0 kubenswrapper[10444]: I1205 10:38:28.458032 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 10:38:28.458296 master-0 kubenswrapper[10444]: I1205 10:38:28.458032 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-pz2dt" Dec 05 10:38:28.458891 master-0 kubenswrapper[10444]: I1205 10:38:28.458860 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 10:38:28.485194 master-0 kubenswrapper[10444]: I1205 10:38:28.485015 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t"] Dec 05 10:38:28.564374 master-0 kubenswrapper[10444]: I1205 10:38:28.564292 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrwt6\" (UniqueName: \"kubernetes.io/projected/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-kube-api-access-lrwt6\") pod \"control-plane-machine-set-operator-7df95c79b5-qnq6t\" (UID: \"24444474-aa9a-4a0f-8b4d-90f0009e0dc7\") " pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:38:28.564553 master-0 kubenswrapper[10444]: I1205 10:38:28.564392 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-7df95c79b5-qnq6t\" (UID: \"24444474-aa9a-4a0f-8b4d-90f0009e0dc7\") " pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:38:28.666641 master-0 kubenswrapper[10444]: I1205 10:38:28.665259 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrwt6\" (UniqueName: \"kubernetes.io/projected/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-kube-api-access-lrwt6\") pod \"control-plane-machine-set-operator-7df95c79b5-qnq6t\" (UID: \"24444474-aa9a-4a0f-8b4d-90f0009e0dc7\") " pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:38:28.666641 master-0 kubenswrapper[10444]: I1205 10:38:28.665535 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-7df95c79b5-qnq6t\" (UID: \"24444474-aa9a-4a0f-8b4d-90f0009e0dc7\") " pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:38:28.669414 master-0 kubenswrapper[10444]: I1205 10:38:28.669156 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-7df95c79b5-qnq6t\" (UID: \"24444474-aa9a-4a0f-8b4d-90f0009e0dc7\") " pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:38:28.695017 master-0 kubenswrapper[10444]: I1205 10:38:28.694958 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrwt6\" (UniqueName: \"kubernetes.io/projected/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-kube-api-access-lrwt6\") pod \"control-plane-machine-set-operator-7df95c79b5-qnq6t\" (UID: \"24444474-aa9a-4a0f-8b4d-90f0009e0dc7\") " pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:38:28.772494 master-0 kubenswrapper[10444]: I1205 10:38:28.770636 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:38:28.894020 master-0 kubenswrapper[10444]: I1205 10:38:28.893994 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler_installer-1-master-0_097d0e8f-46d4-4dfe-862f-54b2b9b21a89/installer/0.log" Dec 05 10:38:28.894178 master-0 kubenswrapper[10444]: I1205 10:38:28.894054 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:38:28.899195 master-0 kubenswrapper[10444]: I1205 10:38:28.899165 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:38:28.985775 master-0 kubenswrapper[10444]: I1205 10:38:28.984866 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kube-api-access\") pod \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " Dec 05 10:38:28.985775 master-0 kubenswrapper[10444]: I1205 10:38:28.985026 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kubelet-dir\") pod \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " Dec 05 10:38:28.985775 master-0 kubenswrapper[10444]: I1205 10:38:28.985121 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-var-lock" (OuterVolumeSpecName: "var-lock") pod "097d0e8f-46d4-4dfe-862f-54b2b9b21a89" (UID: "097d0e8f-46d4-4dfe-862f-54b2b9b21a89"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:38:28.985775 master-0 kubenswrapper[10444]: I1205 10:38:28.985059 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-var-lock\") pod \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\" (UID: \"097d0e8f-46d4-4dfe-862f-54b2b9b21a89\") " Dec 05 10:38:28.987210 master-0 kubenswrapper[10444]: I1205 10:38:28.985992 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:28.987210 master-0 kubenswrapper[10444]: I1205 10:38:28.986977 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "097d0e8f-46d4-4dfe-862f-54b2b9b21a89" (UID: "097d0e8f-46d4-4dfe-862f-54b2b9b21a89"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:38:28.991678 master-0 kubenswrapper[10444]: I1205 10:38:28.991628 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "097d0e8f-46d4-4dfe-862f-54b2b9b21a89" (UID: "097d0e8f-46d4-4dfe-862f-54b2b9b21a89"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:38:29.086594 master-0 kubenswrapper[10444]: I1205 10:38:29.086556 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:29.086787 master-0 kubenswrapper[10444]: I1205 10:38:29.086770 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/097d0e8f-46d4-4dfe-862f-54b2b9b21a89-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:29.257002 master-0 kubenswrapper[10444]: I1205 10:38:29.254734 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t"] Dec 05 10:38:29.268547 master-0 kubenswrapper[10444]: W1205 10:38:29.264509 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24444474_aa9a_4a0f_8b4d_90f0009e0dc7.slice/crio-97d00d6aa4d6661e2be349a4e20636f067fc69912c7110a63d99e820d5970ef4 WatchSource:0}: Error finding container 97d00d6aa4d6661e2be349a4e20636f067fc69912c7110a63d99e820d5970ef4: Status 404 returned error can't find the container with id 97d00d6aa4d6661e2be349a4e20636f067fc69912c7110a63d99e820d5970ef4 Dec 05 10:38:29.419831 master-0 kubenswrapper[10444]: I1205 10:38:29.419782 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler_installer-1-master-0_097d0e8f-46d4-4dfe-862f-54b2b9b21a89/installer/0.log" Dec 05 10:38:29.420054 master-0 kubenswrapper[10444]: I1205 10:38:29.419862 10444 generic.go:334] "Generic (PLEG): container finished" podID="097d0e8f-46d4-4dfe-862f-54b2b9b21a89" containerID="f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c" exitCode=1 Dec 05 10:38:29.420054 master-0 kubenswrapper[10444]: I1205 10:38:29.419920 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-1-master-0" Dec 05 10:38:29.420054 master-0 kubenswrapper[10444]: I1205 10:38:29.419969 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-1-master-0" event={"ID":"097d0e8f-46d4-4dfe-862f-54b2b9b21a89","Type":"ContainerDied","Data":"f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c"} Dec 05 10:38:29.420054 master-0 kubenswrapper[10444]: I1205 10:38:29.420032 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-1-master-0" event={"ID":"097d0e8f-46d4-4dfe-862f-54b2b9b21a89","Type":"ContainerDied","Data":"a16f63e21ba69da6613675e95c4ec003be827b3b94b2ba8c9b38f86425a10d28"} Dec 05 10:38:29.420225 master-0 kubenswrapper[10444]: I1205 10:38:29.420069 10444 scope.go:117] "RemoveContainer" containerID="f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c" Dec 05 10:38:29.421227 master-0 kubenswrapper[10444]: I1205 10:38:29.421200 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" event={"ID":"24444474-aa9a-4a0f-8b4d-90f0009e0dc7","Type":"ContainerStarted","Data":"97d00d6aa4d6661e2be349a4e20636f067fc69912c7110a63d99e820d5970ef4"} Dec 05 10:38:29.433301 master-0 kubenswrapper[10444]: I1205 10:38:29.433255 10444 scope.go:117] "RemoveContainer" containerID="f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c" Dec 05 10:38:29.433768 master-0 kubenswrapper[10444]: E1205 10:38:29.433739 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c\": container with ID starting with f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c not found: ID does not exist" containerID="f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c" Dec 05 10:38:29.433832 master-0 kubenswrapper[10444]: I1205 10:38:29.433770 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c"} err="failed to get container status \"f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c\": rpc error: code = NotFound desc = could not find container \"f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c\": container with ID starting with f16577a9f47e3aaa5a0095dbc65505404a872153cfe4a338661611efddc9c59c not found: ID does not exist" Dec 05 10:38:29.452133 master-0 kubenswrapper[10444]: I1205 10:38:29.452083 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-scheduler/installer-1-master-0"] Dec 05 10:38:29.456812 master-0 kubenswrapper[10444]: I1205 10:38:29.456751 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-scheduler/installer-1-master-0"] Dec 05 10:38:29.599637 master-0 kubenswrapper[10444]: I1205 10:38:29.599578 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="097d0e8f-46d4-4dfe-862f-54b2b9b21a89" path="/var/lib/kubelet/pods/097d0e8f-46d4-4dfe-862f-54b2b9b21a89/volumes" Dec 05 10:38:30.694728 master-0 kubenswrapper[10444]: I1205 10:38:30.694640 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-f797d8546-65t96"] Dec 05 10:38:30.695300 master-0 kubenswrapper[10444]: E1205 10:38:30.695129 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097d0e8f-46d4-4dfe-862f-54b2b9b21a89" containerName="installer" Dec 05 10:38:30.695300 master-0 kubenswrapper[10444]: I1205 10:38:30.695146 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="097d0e8f-46d4-4dfe-862f-54b2b9b21a89" containerName="installer" Dec 05 10:38:30.695390 master-0 kubenswrapper[10444]: I1205 10:38:30.695351 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="097d0e8f-46d4-4dfe-862f-54b2b9b21a89" containerName="installer" Dec 05 10:38:30.696341 master-0 kubenswrapper[10444]: I1205 10:38:30.696316 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.698807 master-0 kubenswrapper[10444]: I1205 10:38:30.698780 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 10:38:30.699021 master-0 kubenswrapper[10444]: I1205 10:38:30.698969 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 10:38:30.699088 master-0 kubenswrapper[10444]: I1205 10:38:30.699027 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 10:38:30.699203 master-0 kubenswrapper[10444]: I1205 10:38:30.699182 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 10:38:30.699309 master-0 kubenswrapper[10444]: I1205 10:38:30.699296 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-2qdv2" Dec 05 10:38:30.702935 master-0 kubenswrapper[10444]: I1205 10:38:30.702902 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 10:38:30.806372 master-0 kubenswrapper[10444]: I1205 10:38:30.806269 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-config\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.806607 master-0 kubenswrapper[10444]: I1205 10:38:30.806443 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt67z\" (UniqueName: \"kubernetes.io/projected/e4c3f514-33e8-4deb-932e-56ddf7281c4c-kube-api-access-nt67z\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.807505 master-0 kubenswrapper[10444]: I1205 10:38:30.807317 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e4c3f514-33e8-4deb-932e-56ddf7281c4c-machine-approver-tls\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.807505 master-0 kubenswrapper[10444]: I1205 10:38:30.807383 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-auth-proxy-config\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.908615 master-0 kubenswrapper[10444]: I1205 10:38:30.908530 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e4c3f514-33e8-4deb-932e-56ddf7281c4c-machine-approver-tls\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.909006 master-0 kubenswrapper[10444]: I1205 10:38:30.908747 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-auth-proxy-config\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.909006 master-0 kubenswrapper[10444]: I1205 10:38:30.908826 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-config\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.909006 master-0 kubenswrapper[10444]: I1205 10:38:30.908875 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt67z\" (UniqueName: \"kubernetes.io/projected/e4c3f514-33e8-4deb-932e-56ddf7281c4c-kube-api-access-nt67z\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.909566 master-0 kubenswrapper[10444]: I1205 10:38:30.909530 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-config\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.909665 master-0 kubenswrapper[10444]: I1205 10:38:30.909621 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-auth-proxy-config\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.912457 master-0 kubenswrapper[10444]: I1205 10:38:30.912391 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e4c3f514-33e8-4deb-932e-56ddf7281c4c-machine-approver-tls\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:30.928211 master-0 kubenswrapper[10444]: I1205 10:38:30.928162 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt67z\" (UniqueName: \"kubernetes.io/projected/e4c3f514-33e8-4deb-932e-56ddf7281c4c-kube-api-access-nt67z\") pod \"machine-approver-f797d8546-65t96\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:31.028925 master-0 kubenswrapper[10444]: I1205 10:38:31.028760 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:38:31.115415 master-0 kubenswrapper[10444]: W1205 10:38:31.115347 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4c3f514_33e8_4deb_932e_56ddf7281c4c.slice/crio-4c18eba53ef1030085a5342e05c911e53f018f33632df043d6b83c7bde56184f WatchSource:0}: Error finding container 4c18eba53ef1030085a5342e05c911e53f018f33632df043d6b83c7bde56184f: Status 404 returned error can't find the container with id 4c18eba53ef1030085a5342e05c911e53f018f33632df043d6b83c7bde56184f Dec 05 10:38:31.440172 master-0 kubenswrapper[10444]: I1205 10:38:31.440108 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" event={"ID":"24444474-aa9a-4a0f-8b4d-90f0009e0dc7","Type":"ContainerStarted","Data":"f67f402cd7e32dc638b6b2ed5376d39de0b7eae17574824302e014030169c159"} Dec 05 10:38:31.442789 master-0 kubenswrapper[10444]: I1205 10:38:31.442742 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" event={"ID":"e4c3f514-33e8-4deb-932e-56ddf7281c4c","Type":"ContainerStarted","Data":"e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608"} Dec 05 10:38:31.442877 master-0 kubenswrapper[10444]: I1205 10:38:31.442793 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" event={"ID":"e4c3f514-33e8-4deb-932e-56ddf7281c4c","Type":"ContainerStarted","Data":"4c18eba53ef1030085a5342e05c911e53f018f33632df043d6b83c7bde56184f"} Dec 05 10:38:31.458039 master-0 kubenswrapper[10444]: I1205 10:38:31.457932 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" podStartSLOduration=1.539407621 podStartE2EDuration="3.45790771s" podCreationTimestamp="2025-12-05 10:38:28 +0000 UTC" firstStartedPulling="2025-12-05 10:38:29.266515056 +0000 UTC m=+49.858427613" lastFinishedPulling="2025-12-05 10:38:31.185015125 +0000 UTC m=+51.776927702" observedRunningTime="2025-12-05 10:38:31.455758269 +0000 UTC m=+52.047670876" watchObservedRunningTime="2025-12-05 10:38:31.45790771 +0000 UTC m=+52.049820317" Dec 05 10:38:32.205185 master-0 kubenswrapper[10444]: I1205 10:38:32.203961 10444 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-etcd/etcd-master-0-master-0"] Dec 05 10:38:32.205185 master-0 kubenswrapper[10444]: I1205 10:38:32.204242 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-etcd/etcd-master-0-master-0" podUID="cc0396a9a2689b3e8c132c12640cbe83" containerName="etcdctl" containerID="cri-o://619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b" gracePeriod=30 Dec 05 10:38:32.211862 master-0 kubenswrapper[10444]: I1205 10:38:32.204294 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-etcd/etcd-master-0-master-0" podUID="cc0396a9a2689b3e8c132c12640cbe83" containerName="etcd" containerID="cri-o://99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c" gracePeriod=30 Dec 05 10:38:32.216037 master-0 kubenswrapper[10444]: I1205 10:38:32.215631 10444 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-master-0"] Dec 05 10:38:32.216037 master-0 kubenswrapper[10444]: E1205 10:38:32.215861 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0396a9a2689b3e8c132c12640cbe83" containerName="etcd" Dec 05 10:38:32.216037 master-0 kubenswrapper[10444]: I1205 10:38:32.215878 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0396a9a2689b3e8c132c12640cbe83" containerName="etcd" Dec 05 10:38:32.216037 master-0 kubenswrapper[10444]: E1205 10:38:32.215888 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc0396a9a2689b3e8c132c12640cbe83" containerName="etcdctl" Dec 05 10:38:32.216037 master-0 kubenswrapper[10444]: I1205 10:38:32.215895 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc0396a9a2689b3e8c132c12640cbe83" containerName="etcdctl" Dec 05 10:38:32.216037 master-0 kubenswrapper[10444]: I1205 10:38:32.215983 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc0396a9a2689b3e8c132c12640cbe83" containerName="etcdctl" Dec 05 10:38:32.216037 master-0 kubenswrapper[10444]: I1205 10:38:32.215998 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc0396a9a2689b3e8c132c12640cbe83" containerName="etcd" Dec 05 10:38:32.217529 master-0 kubenswrapper[10444]: I1205 10:38:32.217495 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.235873 master-0 kubenswrapper[10444]: I1205 10:38:32.228643 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-log-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.235873 master-0 kubenswrapper[10444]: I1205 10:38:32.228812 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-cert-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.235873 master-0 kubenswrapper[10444]: I1205 10:38:32.228896 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-resource-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.235873 master-0 kubenswrapper[10444]: I1205 10:38:32.228985 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-static-pod-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.235873 master-0 kubenswrapper[10444]: I1205 10:38:32.229071 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-data-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.235873 master-0 kubenswrapper[10444]: I1205 10:38:32.229121 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-usr-local-bin\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.330869 master-0 kubenswrapper[10444]: I1205 10:38:32.330629 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-static-pod-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.330869 master-0 kubenswrapper[10444]: I1205 10:38:32.330713 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-data-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.330869 master-0 kubenswrapper[10444]: I1205 10:38:32.330782 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-usr-local-bin\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.330869 master-0 kubenswrapper[10444]: I1205 10:38:32.330805 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-log-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.330869 master-0 kubenswrapper[10444]: I1205 10:38:32.330807 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-static-pod-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.330869 master-0 kubenswrapper[10444]: I1205 10:38:32.330836 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-cert-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.330869 master-0 kubenswrapper[10444]: I1205 10:38:32.330900 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-usr-local-bin\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.330869 master-0 kubenswrapper[10444]: I1205 10:38:32.330922 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-data-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.331724 master-0 kubenswrapper[10444]: I1205 10:38:32.330928 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-log-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.331724 master-0 kubenswrapper[10444]: I1205 10:38:32.331098 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-cert-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.332008 master-0 kubenswrapper[10444]: I1205 10:38:32.331951 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-resource-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:32.332008 master-0 kubenswrapper[10444]: I1205 10:38:32.331985 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-resource-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:38:33.456307 master-0 kubenswrapper[10444]: I1205 10:38:33.456084 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" event={"ID":"e4c3f514-33e8-4deb-932e-56ddf7281c4c","Type":"ContainerStarted","Data":"764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d"} Dec 05 10:38:45.280382 master-0 kubenswrapper[10444]: E1205 10:38:45.280328 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="Internal error occurred: admission plugin \"LimitRanger\" failed to complete mutation in 13s" pod="openshift-etcd/etcd-master-0" Dec 05 10:38:45.280996 master-0 kubenswrapper[10444]: I1205 10:38:45.280808 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-master-0" Dec 05 10:38:45.293515 master-0 kubenswrapper[10444]: W1205 10:38:45.293401 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc24e01603234fe8003f8aae8171b0065.slice/crio-9e8deb0a79a04a088af836af85742626a8778be7f61d22448e3a0b846ff8b05a WatchSource:0}: Error finding container 9e8deb0a79a04a088af836af85742626a8778be7f61d22448e3a0b846ff8b05a: Status 404 returned error can't find the container with id 9e8deb0a79a04a088af836af85742626a8778be7f61d22448e3a0b846ff8b05a Dec 05 10:38:45.513151 master-0 kubenswrapper[10444]: I1205 10:38:45.513100 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"9e8deb0a79a04a088af836af85742626a8778be7f61d22448e3a0b846ff8b05a"} Dec 05 10:38:46.446341 master-0 kubenswrapper[10444]: I1205 10:38:46.446257 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:38:46.521911 master-0 kubenswrapper[10444]: I1205 10:38:46.521852 10444 generic.go:334] "Generic (PLEG): container finished" podID="8b47694fcc32464ab24d09c23d6efb57" containerID="303d7cff916f69b65cb8e8b72a47e944bb182f86fc3e9bfde505fa8e541eb3b6" exitCode=1 Dec 05 10:38:46.521911 master-0 kubenswrapper[10444]: I1205 10:38:46.521913 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerDied","Data":"303d7cff916f69b65cb8e8b72a47e944bb182f86fc3e9bfde505fa8e541eb3b6"} Dec 05 10:38:46.522646 master-0 kubenswrapper[10444]: I1205 10:38:46.521951 10444 scope.go:117] "RemoveContainer" containerID="6c524eb37d9eb519540115909748ea1e0272427fa38212f1b6de731f3f942edd" Dec 05 10:38:46.522646 master-0 kubenswrapper[10444]: I1205 10:38:46.522507 10444 scope.go:117] "RemoveContainer" containerID="303d7cff916f69b65cb8e8b72a47e944bb182f86fc3e9bfde505fa8e541eb3b6" Dec 05 10:38:46.525250 master-0 kubenswrapper[10444]: I1205 10:38:46.524869 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2"} Dec 05 10:38:46.526455 master-0 kubenswrapper[10444]: I1205 10:38:46.526399 10444 generic.go:334] "Generic (PLEG): container finished" podID="e149cbe7-1ec3-4a06-af61-3a8906b8e9ef" containerID="86721e833ddfe5b80f087deb054f43ae1d6033b98f09d3f999eb1efccfb11215" exitCode=0 Dec 05 10:38:46.526603 master-0 kubenswrapper[10444]: I1205 10:38:46.526460 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-1-master-0" event={"ID":"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef","Type":"ContainerDied","Data":"86721e833ddfe5b80f087deb054f43ae1d6033b98f09d3f999eb1efccfb11215"} Dec 05 10:38:47.533768 master-0 kubenswrapper[10444]: I1205 10:38:47.533706 10444 generic.go:334] "Generic (PLEG): container finished" podID="c24e01603234fe8003f8aae8171b0065" containerID="f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2" exitCode=0 Dec 05 10:38:47.534444 master-0 kubenswrapper[10444]: I1205 10:38:47.533766 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerDied","Data":"f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2"} Dec 05 10:38:47.537300 master-0 kubenswrapper[10444]: I1205 10:38:47.537227 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerStarted","Data":"e7ce8dc7bd304b25a361ecabc9d8faa1f0ae1eca9ef1f9dc24bcebc4d09b4a20"} Dec 05 10:38:47.815966 master-0 kubenswrapper[10444]: I1205 10:38:47.815900 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-1-master-0" Dec 05 10:38:48.013066 master-0 kubenswrapper[10444]: I1205 10:38:48.012987 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kubelet-dir\") pod \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " Dec 05 10:38:48.013066 master-0 kubenswrapper[10444]: I1205 10:38:48.013047 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-var-lock\") pod \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " Dec 05 10:38:48.013301 master-0 kubenswrapper[10444]: I1205 10:38:48.013097 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kube-api-access\") pod \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\" (UID: \"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef\") " Dec 05 10:38:48.013501 master-0 kubenswrapper[10444]: I1205 10:38:48.013415 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-var-lock" (OuterVolumeSpecName: "var-lock") pod "e149cbe7-1ec3-4a06-af61-3a8906b8e9ef" (UID: "e149cbe7-1ec3-4a06-af61-3a8906b8e9ef"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:38:48.013578 master-0 kubenswrapper[10444]: I1205 10:38:48.013543 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e149cbe7-1ec3-4a06-af61-3a8906b8e9ef" (UID: "e149cbe7-1ec3-4a06-af61-3a8906b8e9ef"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:38:48.017309 master-0 kubenswrapper[10444]: I1205 10:38:48.017237 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e149cbe7-1ec3-4a06-af61-3a8906b8e9ef" (UID: "e149cbe7-1ec3-4a06-af61-3a8906b8e9ef"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:38:48.067551 master-0 kubenswrapper[10444]: I1205 10:38:48.067306 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:38:48.114008 master-0 kubenswrapper[10444]: I1205 10:38:48.113932 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:48.114008 master-0 kubenswrapper[10444]: I1205 10:38:48.113973 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:48.114008 master-0 kubenswrapper[10444]: I1205 10:38:48.113984 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e149cbe7-1ec3-4a06-af61-3a8906b8e9ef-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:38:48.543486 master-0 kubenswrapper[10444]: I1205 10:38:48.543342 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-1-master-0" event={"ID":"e149cbe7-1ec3-4a06-af61-3a8906b8e9ef","Type":"ContainerDied","Data":"a87c5cb2ce0ea32814514a0a2632b27b31a950e85396e1a0393057f93ec2e154"} Dec 05 10:38:48.543486 master-0 kubenswrapper[10444]: I1205 10:38:48.543455 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a87c5cb2ce0ea32814514a0a2632b27b31a950e85396e1a0393057f93ec2e154" Dec 05 10:38:48.543486 master-0 kubenswrapper[10444]: I1205 10:38:48.543368 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-1-master-0" Dec 05 10:38:48.544918 master-0 kubenswrapper[10444]: I1205 10:38:48.544849 10444 generic.go:334] "Generic (PLEG): container finished" podID="5e09e2af7200e6f9be469dbfd9bb1127" containerID="f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782" exitCode=1 Dec 05 10:38:48.545041 master-0 kubenswrapper[10444]: I1205 10:38:48.544975 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-scheduler-master-0" event={"ID":"5e09e2af7200e6f9be469dbfd9bb1127","Type":"ContainerDied","Data":"f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782"} Dec 05 10:38:48.545610 master-0 kubenswrapper[10444]: I1205 10:38:48.545565 10444 scope.go:117] "RemoveContainer" containerID="f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782" Dec 05 10:38:49.551537 master-0 kubenswrapper[10444]: I1205 10:38:49.551482 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-scheduler-master-0" event={"ID":"5e09e2af7200e6f9be469dbfd9bb1127","Type":"ContainerStarted","Data":"2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000"} Dec 05 10:38:50.257454 master-0 kubenswrapper[10444]: E1205 10:38:50.257325 10444 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:38:50.727209 master-0 kubenswrapper[10444]: I1205 10:38:50.727143 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:38:52.568804 master-0 kubenswrapper[10444]: I1205 10:38:52.568765 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5_1e69ce9e-4e6f-4015-9ba6-5a7942570190/openshift-controller-manager-operator/0.log" Dec 05 10:38:52.569309 master-0 kubenswrapper[10444]: I1205 10:38:52.568815 10444 generic.go:334] "Generic (PLEG): container finished" podID="1e69ce9e-4e6f-4015-9ba6-5a7942570190" containerID="41f047cf59b8b167740e4f4b07cc948566d4077a26e6c889e254fc22e5a5bd11" exitCode=1 Dec 05 10:38:52.569309 master-0 kubenswrapper[10444]: I1205 10:38:52.568843 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" event={"ID":"1e69ce9e-4e6f-4015-9ba6-5a7942570190","Type":"ContainerDied","Data":"41f047cf59b8b167740e4f4b07cc948566d4077a26e6c889e254fc22e5a5bd11"} Dec 05 10:38:52.569309 master-0 kubenswrapper[10444]: I1205 10:38:52.569191 10444 scope.go:117] "RemoveContainer" containerID="41f047cf59b8b167740e4f4b07cc948566d4077a26e6c889e254fc22e5a5bd11" Dec 05 10:38:53.575665 master-0 kubenswrapper[10444]: I1205 10:38:53.575602 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5_1e69ce9e-4e6f-4015-9ba6-5a7942570190/openshift-controller-manager-operator/0.log" Dec 05 10:38:53.575665 master-0 kubenswrapper[10444]: I1205 10:38:53.575664 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" event={"ID":"1e69ce9e-4e6f-4015-9ba6-5a7942570190","Type":"ContainerStarted","Data":"ded2f19eb6b7e89481ac839fb3ba5dd47ee60cff52debe8dba17401c93d0ac1a"} Dec 05 10:38:58.067772 master-0 kubenswrapper[10444]: I1205 10:38:58.067710 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:38:59.602737 master-0 kubenswrapper[10444]: I1205 10:38:59.602678 10444 generic.go:334] "Generic (PLEG): container finished" podID="cc0396a9a2689b3e8c132c12640cbe83" containerID="99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c" exitCode=0 Dec 05 10:39:00.258076 master-0 kubenswrapper[10444]: E1205 10:39:00.257991 10444 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:39:00.542892 master-0 kubenswrapper[10444]: E1205 10:39:00.542764 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="Internal error occurred: admission plugin \"LimitRanger\" failed to complete mutation in 13s" pod="openshift-etcd/etcd-master-0" Dec 05 10:39:01.068793 master-0 kubenswrapper[10444]: I1205 10:39:01.068698 10444 prober.go:107] "Probe failed" probeType="Startup" pod="kube-system/bootstrap-kube-controller-manager-master-0" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.32.10:10257/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 10:39:01.392831 master-0 kubenswrapper[10444]: E1205 10:39:01.392566 10444 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:38:51Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:38:51Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:38:51Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:38:51Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b\\\"],\\\"sizeBytes\\\":1631758507},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9014f384de5f9a0b7418d5869ad349abb9588d16bd09ed650a163c045315dbff\\\"],\\\"sizeBytes\\\":1232140918},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6\\\"],\\\"sizeBytes\\\":983705650},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\\\"],\\\"sizeBytes\\\":938303566},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:631a3798b749fecc041a99929eb946618df723e15055e805ff752a1a1273481c\\\"],\\\"sizeBytes\\\":870567329},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\\\"],\\\"sizeBytes\\\":857069957},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b12f830c3316aa4dc061c2d00c74126282b3e2bcccc301eab00d57fff3c4c7c\\\"],\\\"sizeBytes\\\":767284906},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cb3ec61f9a932a9ad13bdeb44bcf9477a8d5f728151d7f19ed3ef7d4b02b3a82\\\"],\\\"sizeBytes\\\":682371258},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:916566bb9d0143352324233d460ad94697719c11c8c9158e3aea8f475941751f\\\"],\\\"sizeBytes\\\":677523572},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b\\\"],\\\"sizeBytes\\\":672407260},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9724d2036305cbd729e1f484c5bad89971de977fff8a6723fef1873858dd1123\\\"],\\\"sizeBytes\\\":616108962},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:df606f3b71d4376d1a2108c09f0d3dab455fc30bcb67c60e91590c105e9025bf\\\"],\\\"sizeBytes\\\":583836304},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2\\\"],\\\"sizeBytes\\\":576619763},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8\\\"],\\\"sizeBytes\\\":552673986},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc\\\"],\\\"sizeBytes\\\":543227406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718\\\"],\\\"sizeBytes\\\":532719167},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cfde59e48cd5dee3721f34d249cb119cc3259fd857965d34f9c7ed83b0c363a1\\\"],\\\"sizeBytes\\\":532402162},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f4d4282cb53325e737ad68abbfcb70687ae04fb50353f4f0ba0ba5703b15009a\\\"],\\\"sizeBytes\\\":512838054},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\\\"],\\\"sizeBytes\\\":512452153},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660\\\"],\\\"sizeBytes\\\":509437356},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df\\\"],\\\"sizeBytes\\\":507687221},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8375671da86aa527ee7e291d86971b0baa823ffc7663b5a983084456e76c0f59\\\"],\\\"sizeBytes\\\":506741476},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b\\\"],\\\"sizeBytes\\\":505649178},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:86af77350cfe6fd69280157e4162aa0147873d9431c641ae4ad3e881ff768a73\\\"],\\\"sizeBytes\\\":505628211},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\\\"],\\\"sizeBytes\\\":503340749},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8139ed65c0a0a4b0f253b715c11cc52be027efe8a4774da9ccce35c78ef439da\\\"],\\\"sizeBytes\\\":503011144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4\\\"],\\\"sizeBytes\\\":502436444},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10e57ca7611f79710f05777dc6a8f31c7e04eb09da4d8d793a5acfbf0e4692d7\\\"],\\\"sizeBytes\\\":500943492},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce\\\"],\\\"sizeBytes\\\":500848684},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c\\\"],\\\"sizeBytes\\\":499798563},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96\\\"],\\\"sizeBytes\\\":499705918},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:75d996f6147edb88c09fd1a052099de66638590d7d03a735006244bc9e19f898\\\"],\\\"sizeBytes\\\":499082775},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3b8d91a25eeb9f02041e947adb3487da3e7ab8449d3d2ad015827e7954df7b34\\\"],\\\"sizeBytes\\\":490455952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68\\\"],\\\"sizeBytes\\\":489528665},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1\\\"],\\\"sizeBytes\\\":481559117},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3a77aa4d03b89ea284e3467a268e5989a77a2ef63e685eb1d5c5ea5b3922b7a\\\"],\\\"sizeBytes\\\":478917802},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eb928c13a46d3fb45f4a881892d023a92d610a5430be0ffd916aaf8da8e7d297\\\"],\\\"sizeBytes\\\":478642572},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fd3e9f8f00a59bda7483ec7dc8a0ed602f9ca30e3d72b22072dbdf2819da3f61\\\"],\\\"sizeBytes\\\":465144618},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3c1edf52f70bf9b1d1457e0c4111bc79cdaa1edd659ddbdb9d8176eff8b46956\\\"],\\\"sizeBytes\\\":462727837},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8cc27777e72233024fe84ee1faa168aec715a0b24912a3ce70715ddccba328df\\\"],\\\"sizeBytes\\\":461702648},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\\\"],\\\"sizeBytes\\\":459552216},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d3ce2cbf1032ad0f24f204db73687002fcf302e86ebde3945801c74351b64576\\\"],\\\"sizeBytes\\\":458169255},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7664a2d4cb10e82ed32abbf95799f43fc3d10135d7dd94799730de504a89680a\\\"],\\\"sizeBytes\\\":452589750},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4ecc5bac651ff1942865baee5159582e9602c89b47eeab18400a32abcba8f690\\\"],\\\"sizeBytes\\\":451039520},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2632d7f05d5a992e91038ded81c715898f3fe803420a9b67a0201e9fd8075213\\\"],\\\"sizeBytes\\\":443291941},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3f0aa9cd04713acc5c6fea721bd849e1500da8ae945e0b32000887f34d786e0b\\\"],\\\"sizeBytes\\\":442509555},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e438b814f8e16f00b3fc4b69991af80eee79ae111d2a707f34aa64b2ccbb6eb\\\"],\\\"sizeBytes\\\":437737925},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b00c658332d6c6786bd969b26097c20a78c79c045f1692a8809234f5fb586c22\\\"],\\\"sizeBytes\\\":433122306},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a3d37aa7a22c68afa963ecfb4b43c52cccf152580cd66e4d5382fb69e4037cc\\\"],\\\"sizeBytes\\\":406053031},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9432c13d76bd4ba4eb9197c050cf88c0d701fa2055eeb59257e2e23901f9fdff\\\"],\\\"sizeBytes\\\":401810450}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"master-0\": Patch \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0/status?timeout=10s\": context deadline exceeded" Dec 05 10:39:01.616181 master-0 kubenswrapper[10444]: I1205 10:39:01.616086 10444 generic.go:334] "Generic (PLEG): container finished" podID="c24e01603234fe8003f8aae8171b0065" containerID="d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc" exitCode=0 Dec 05 10:39:01.616181 master-0 kubenswrapper[10444]: I1205 10:39:01.616139 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerDied","Data":"d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc"} Dec 05 10:39:02.329468 master-0 kubenswrapper[10444]: I1205 10:39:02.329392 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0-master-0_cc0396a9a2689b3e8c132c12640cbe83/etcdctl/0.log" Dec 05 10:39:02.329943 master-0 kubenswrapper[10444]: I1205 10:39:02.329524 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:39:02.472895 master-0 kubenswrapper[10444]: I1205 10:39:02.472817 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-data-dir\") pod \"cc0396a9a2689b3e8c132c12640cbe83\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " Dec 05 10:39:02.473142 master-0 kubenswrapper[10444]: I1205 10:39:02.472966 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-certs\") pod \"cc0396a9a2689b3e8c132c12640cbe83\" (UID: \"cc0396a9a2689b3e8c132c12640cbe83\") " Dec 05 10:39:02.473142 master-0 kubenswrapper[10444]: I1205 10:39:02.472965 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-data-dir" (OuterVolumeSpecName: "data-dir") pod "cc0396a9a2689b3e8c132c12640cbe83" (UID: "cc0396a9a2689b3e8c132c12640cbe83"). InnerVolumeSpecName "data-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:39:02.473216 master-0 kubenswrapper[10444]: I1205 10:39:02.473139 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-certs" (OuterVolumeSpecName: "certs") pod "cc0396a9a2689b3e8c132c12640cbe83" (UID: "cc0396a9a2689b3e8c132c12640cbe83"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:39:02.473576 master-0 kubenswrapper[10444]: I1205 10:39:02.473541 10444 reconciler_common.go:293] "Volume detached for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-data-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:39:02.473612 master-0 kubenswrapper[10444]: I1205 10:39:02.473578 10444 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/host-path/cc0396a9a2689b3e8c132c12640cbe83-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 10:39:02.623780 master-0 kubenswrapper[10444]: I1205 10:39:02.623715 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0-master-0_cc0396a9a2689b3e8c132c12640cbe83/etcdctl/0.log" Dec 05 10:39:02.624037 master-0 kubenswrapper[10444]: I1205 10:39:02.623807 10444 generic.go:334] "Generic (PLEG): container finished" podID="cc0396a9a2689b3e8c132c12640cbe83" containerID="619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b" exitCode=137 Dec 05 10:39:02.624037 master-0 kubenswrapper[10444]: I1205 10:39:02.623882 10444 scope.go:117] "RemoveContainer" containerID="99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c" Dec 05 10:39:02.624037 master-0 kubenswrapper[10444]: I1205 10:39:02.623938 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:39:02.638761 master-0 kubenswrapper[10444]: I1205 10:39:02.638724 10444 scope.go:117] "RemoveContainer" containerID="619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b" Dec 05 10:39:02.655859 master-0 kubenswrapper[10444]: I1205 10:39:02.655809 10444 scope.go:117] "RemoveContainer" containerID="99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c" Dec 05 10:39:02.656455 master-0 kubenswrapper[10444]: E1205 10:39:02.656406 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c\": container with ID starting with 99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c not found: ID does not exist" containerID="99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c" Dec 05 10:39:02.656537 master-0 kubenswrapper[10444]: I1205 10:39:02.656464 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c"} err="failed to get container status \"99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c\": rpc error: code = NotFound desc = could not find container \"99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c\": container with ID starting with 99d0e79d7fd1b2c314e10bf397b150542ea03baf684ef88cc3a019de3638643c not found: ID does not exist" Dec 05 10:39:02.656537 master-0 kubenswrapper[10444]: I1205 10:39:02.656491 10444 scope.go:117] "RemoveContainer" containerID="619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b" Dec 05 10:39:02.657145 master-0 kubenswrapper[10444]: E1205 10:39:02.657104 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b\": container with ID starting with 619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b not found: ID does not exist" containerID="619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b" Dec 05 10:39:02.657213 master-0 kubenswrapper[10444]: I1205 10:39:02.657146 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b"} err="failed to get container status \"619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b\": rpc error: code = NotFound desc = could not find container \"619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b\": container with ID starting with 619f1638d1b2c9e17eeab6d2bfd76ff6bf2769ba6f88076510e2e24c42a5249b not found: ID does not exist" Dec 05 10:39:03.603183 master-0 kubenswrapper[10444]: I1205 10:39:03.603114 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc0396a9a2689b3e8c132c12640cbe83" path="/var/lib/kubelet/pods/cc0396a9a2689b3e8c132c12640cbe83/volumes" Dec 05 10:39:03.603877 master-0 kubenswrapper[10444]: I1205 10:39:03.603460 10444 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0-master-0" podUID="" Dec 05 10:39:06.652490 master-0 kubenswrapper[10444]: I1205 10:39:06.652408 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-1-master-0_5d8c854f-31f8-42d3-b28c-82c7d1b14ee9/installer/0.log" Dec 05 10:39:06.652490 master-0 kubenswrapper[10444]: I1205 10:39:06.652489 10444 generic.go:334] "Generic (PLEG): container finished" podID="5d8c854f-31f8-42d3-b28c-82c7d1b14ee9" containerID="0d4c12890814d7bd72e2d33d03a6131a3d52c0e2122d3701215dc96f26a231b2" exitCode=1 Dec 05 10:39:06.829375 master-0 kubenswrapper[10444]: E1205 10:39:06.829171 10444 event.go:359] "Server rejected event (will not retry!)" err="Timeout: request did not complete within requested timeout - context deadline exceeded" event="&Event{ObjectMeta:{machine-approver-f797d8546-65t96.187e4b7f0107ae55 openshift-cluster-machine-approver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-cluster-machine-approver,Name:machine-approver-f797d8546-65t96,UID:e4c3f514-33e8-4deb-932e-56ddf7281c4c,APIVersion:v1,ResourceVersion:8584,FieldPath:spec.containers{machine-approver-controller},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8cc27777e72233024fe84ee1faa168aec715a0b24912a3ce70715ddccba328df\" in 1.478s (1.478s including waiting). Image size: 461702648 bytes.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:38:32.826023509 +0000 UTC m=+53.417936076,LastTimestamp:2025-12-05 10:38:32.826023509 +0000 UTC m=+53.417936076,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:39:08.668030 master-0 kubenswrapper[10444]: I1205 10:39:08.667949 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_installer-1-master-0_738d9fe7-2d08-4aa4-842b-9cbeddf120fb/installer/0.log" Dec 05 10:39:08.668030 master-0 kubenswrapper[10444]: I1205 10:39:08.667998 10444 generic.go:334] "Generic (PLEG): container finished" podID="738d9fe7-2d08-4aa4-842b-9cbeddf120fb" containerID="8a8e52c1bfb7c50d31dbfbb099e72a94f8cf7f5c9eac30d42a9e7bb6f30721bf" exitCode=1 Dec 05 10:39:10.259554 master-0 kubenswrapper[10444]: E1205 10:39:10.259402 10444 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:39:10.803697 master-0 kubenswrapper[10444]: I1205 10:39:10.803615 10444 patch_prober.go:28] interesting pod/authentication-operator-6c968fdfdf-t7sl8 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.128.0.23:8443/healthz\": dial tcp 10.128.0.23:8443: connect: connection refused" start-of-body= Dec 05 10:39:10.803938 master-0 kubenswrapper[10444]: I1205 10:39:10.803687 10444 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" podUID="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.128.0.23:8443/healthz\": dial tcp 10.128.0.23:8443: connect: connection refused" Dec 05 10:39:11.068747 master-0 kubenswrapper[10444]: I1205 10:39:11.068534 10444 prober.go:107] "Probe failed" probeType="Startup" pod="kube-system/bootstrap-kube-controller-manager-master-0" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.32.10:10257/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 10:39:11.393961 master-0 kubenswrapper[10444]: E1205 10:39:11.393692 10444 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:39:14.624242 master-0 kubenswrapper[10444]: E1205 10:39:14.624131 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="Internal error occurred: admission plugin \"LimitRanger\" failed to complete mutation in 13s" pod="openshift-etcd/etcd-master-0" Dec 05 10:39:15.710360 master-0 kubenswrapper[10444]: I1205 10:39:15.710253 10444 generic.go:334] "Generic (PLEG): container finished" podID="c24e01603234fe8003f8aae8171b0065" containerID="41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2" exitCode=0 Dec 05 10:39:16.716667 master-0 kubenswrapper[10444]: I1205 10:39:16.716575 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler_installer-3-master-0_6325f9a2-86d5-4fac-9e71-331b2628f965/installer/0.log" Dec 05 10:39:16.716667 master-0 kubenswrapper[10444]: I1205 10:39:16.716632 10444 generic.go:334] "Generic (PLEG): container finished" podID="6325f9a2-86d5-4fac-9e71-331b2628f965" containerID="df5bdcc899d9a7c7268a300825923259315bb7a02282ed76b8b7644641b06995" exitCode=1 Dec 05 10:39:20.260397 master-0 kubenswrapper[10444]: E1205 10:39:20.260168 10444 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:39:20.741597 master-0 kubenswrapper[10444]: I1205 10:39:20.741522 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-node-identity_network-node-identity-ql7j7_d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/approver/0.log" Dec 05 10:39:20.741942 master-0 kubenswrapper[10444]: I1205 10:39:20.741898 10444 generic.go:334] "Generic (PLEG): container finished" podID="d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e" containerID="5674d03e599731519f29f243f076ae28159025706182191e04e4bbfef0819512" exitCode=1 Dec 05 10:39:20.802598 master-0 kubenswrapper[10444]: I1205 10:39:20.802523 10444 patch_prober.go:28] interesting pod/authentication-operator-6c968fdfdf-t7sl8 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.128.0.23:8443/healthz\": dial tcp 10.128.0.23:8443: connect: connection refused" start-of-body= Dec 05 10:39:20.802598 master-0 kubenswrapper[10444]: I1205 10:39:20.802597 10444 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" podUID="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.128.0.23:8443/healthz\": dial tcp 10.128.0.23:8443: connect: connection refused" Dec 05 10:39:21.067821 master-0 kubenswrapper[10444]: I1205 10:39:21.067699 10444 prober.go:107] "Probe failed" probeType="Startup" pod="kube-system/bootstrap-kube-controller-manager-master-0" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.32.10:10257/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 10:39:21.395134 master-0 kubenswrapper[10444]: E1205 10:39:21.394927 10444 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:39:26.780486 master-0 kubenswrapper[10444]: I1205 10:39:26.780228 10444 generic.go:334] "Generic (PLEG): container finished" podID="4825316a-ea9f-4d3d-838b-fa809a6e49c7" containerID="f2785fc5d00b92168028cf4e3273a28c6085e8db64c579bdce5eca572512c7d9" exitCode=0 Dec 05 10:39:26.782415 master-0 kubenswrapper[10444]: I1205 10:39:26.782356 10444 generic.go:334] "Generic (PLEG): container finished" podID="f7a08359-0379-4364-8b0c-ddb58ff605f4" containerID="7f719d929cd8f822ace129247877e00bae5a9457d73d9da2566b48784c2ee8fc" exitCode=0 Dec 05 10:39:28.717140 master-0 kubenswrapper[10444]: E1205 10:39:28.717022 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="Internal error occurred: admission plugin \"LimitRanger\" failed to complete mutation in 13s" pod="openshift-etcd/etcd-master-0" Dec 05 10:39:30.261811 master-0 kubenswrapper[10444]: E1205 10:39:30.261417 10444 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:39:30.261811 master-0 kubenswrapper[10444]: I1205 10:39:30.261514 10444 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 05 10:39:30.803353 master-0 kubenswrapper[10444]: I1205 10:39:30.803295 10444 patch_prober.go:28] interesting pod/authentication-operator-6c968fdfdf-t7sl8 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.128.0.23:8443/healthz\": dial tcp 10.128.0.23:8443: connect: connection refused" start-of-body= Dec 05 10:39:30.803761 master-0 kubenswrapper[10444]: I1205 10:39:30.803722 10444 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" podUID="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.128.0.23:8443/healthz\": dial tcp 10.128.0.23:8443: connect: connection refused" Dec 05 10:39:31.396125 master-0 kubenswrapper[10444]: E1205 10:39:31.396053 10444 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:39:31.818259 master-0 kubenswrapper[10444]: I1205 10:39:31.818189 10444 generic.go:334] "Generic (PLEG): container finished" podID="6f76d12f-5406-47e2-8337-2f50e35376d6" containerID="cc8433167039cff15f2a8553e5790127434ae4c5286a5db0777d417011f80f8c" exitCode=0 Dec 05 10:39:32.824579 master-0 kubenswrapper[10444]: I1205 10:39:32.824465 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_network-operator-79767b7ff9-t8j2j_1a0235af-2cf2-4ad4-b419-764fb56a0107/network-operator/0.log" Dec 05 10:39:32.824579 master-0 kubenswrapper[10444]: I1205 10:39:32.824561 10444 generic.go:334] "Generic (PLEG): container finished" podID="1a0235af-2cf2-4ad4-b419-764fb56a0107" containerID="f1e6d80e2ae7d337901560dde2050e55f54f6b182ab45a5e1cce9ebc5f4043a9" exitCode=255 Dec 05 10:39:33.459524 master-0 kubenswrapper[10444]: I1205 10:39:33.459452 10444 status_manager.go:851] "Failed to get status for pod" podUID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" err="the server was unable to return a response in the time allotted, but may still be processing the request (get pods machine-approver-f797d8546-65t96)" Dec 05 10:39:34.718241 master-0 kubenswrapper[10444]: I1205 10:39:34.718128 10444 patch_prober.go:28] interesting pod/etcd-operator-5bf4d88c6f-n8t5c container/etcd-operator namespace/openshift-etcd-operator: Liveness probe status=failure output="Get \"https://10.128.0.15:8443/healthz\": dial tcp 10.128.0.15:8443: connect: connection refused" start-of-body= Dec 05 10:39:34.718862 master-0 kubenswrapper[10444]: I1205 10:39:34.718255 10444 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" podUID="f7a08359-0379-4364-8b0c-ddb58ff605f4" containerName="etcd-operator" probeResult="failure" output="Get \"https://10.128.0.15:8443/healthz\": dial tcp 10.128.0.15:8443: connect: connection refused" Dec 05 10:39:37.606542 master-0 kubenswrapper[10444]: E1205 10:39:37.606467 10444 mirror_client.go:138] "Failed deleting a mirror pod" err="Timeout: request did not complete within requested timeout - context deadline exceeded" pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:39:37.607673 master-0 kubenswrapper[10444]: E1205 10:39:37.606641 10444 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="34.013s" Dec 05 10:39:37.607673 master-0 kubenswrapper[10444]: I1205 10:39:37.606664 10444 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:39:37.607673 master-0 kubenswrapper[10444]: I1205 10:39:37.607532 10444 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="authentication-operator" containerStatusID={"Type":"cri-o","ID":"dad5a560aae0ea7cbefd19607cf4263a0d5cfc63a06ebe4360483bee7047a720"} pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" containerMessage="Container authentication-operator failed liveness probe, will be restarted" Dec 05 10:39:37.607673 master-0 kubenswrapper[10444]: I1205 10:39:37.607611 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" podUID="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" containerName="authentication-operator" containerID="cri-o://dad5a560aae0ea7cbefd19607cf4263a0d5cfc63a06ebe4360483bee7047a720" gracePeriod=30 Dec 05 10:39:37.615099 master-0 kubenswrapper[10444]: I1205 10:39:37.615041 10444 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0-master-0" podUID="" Dec 05 10:39:38.891252 master-0 kubenswrapper[10444]: I1205 10:39:38.891185 10444 generic.go:334] "Generic (PLEG): container finished" podID="444f8808-e454-4015-9e20-429e715a08c7" containerID="a4e599b9d9dac0eea61bdc89656e6a7ee18f83aa3fd9ba9340f57da02bd4a39b" exitCode=0 Dec 05 10:39:40.261871 master-0 kubenswrapper[10444]: E1205 10:39:40.261738 10444 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="200ms" Dec 05 10:39:40.831399 master-0 kubenswrapper[10444]: E1205 10:39:40.831275 10444 event.go:359] "Server rejected event (will not retry!)" err="Timeout: request did not complete within requested timeout - context deadline exceeded" event="&Event{ObjectMeta:{machine-approver-f797d8546-65t96.187e4b7f102bb3a0 openshift-cluster-machine-approver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-cluster-machine-approver,Name:machine-approver-f797d8546-65t96,UID:e4c3f514-33e8-4deb-932e-56ddf7281c4c,APIVersion:v1,ResourceVersion:8584,FieldPath:spec.containers{machine-approver-controller},},Reason:Created,Message:Created container: machine-approver-controller,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:38:33.0800424 +0000 UTC m=+53.671955007,LastTimestamp:2025-12-05 10:38:33.0800424 +0000 UTC m=+53.671955007,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:39:41.398141 master-0 kubenswrapper[10444]: E1205 10:39:41.398070 10444 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:39:41.398141 master-0 kubenswrapper[10444]: E1205 10:39:41.398124 10444 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 10:39:41.910588 master-0 kubenswrapper[10444]: I1205 10:39:41.910506 10444 generic.go:334] "Generic (PLEG): container finished" podID="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" containerID="dad5a560aae0ea7cbefd19607cf4263a0d5cfc63a06ebe4360483bee7047a720" exitCode=0 Dec 05 10:39:47.947677 master-0 kubenswrapper[10444]: I1205 10:39:47.947599 10444 generic.go:334] "Generic (PLEG): container finished" podID="8b47694fcc32464ab24d09c23d6efb57" containerID="e7ce8dc7bd304b25a361ecabc9d8faa1f0ae1eca9ef1f9dc24bcebc4d09b4a20" exitCode=1 Dec 05 10:39:50.462722 master-0 kubenswrapper[10444]: E1205 10:39:50.462618 10444 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="400ms" Dec 05 10:40:00.864761 master-0 kubenswrapper[10444]: E1205 10:40:00.864678 10444 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": context deadline exceeded" interval="800ms" Dec 05 10:40:01.418760 master-0 kubenswrapper[10444]: E1205 10:40:01.418578 10444 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:39:51Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:39:51Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:39:51Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:39:51Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b\\\"],\\\"sizeBytes\\\":1631758507},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9014f384de5f9a0b7418d5869ad349abb9588d16bd09ed650a163c045315dbff\\\"],\\\"sizeBytes\\\":1232140918},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6\\\"],\\\"sizeBytes\\\":983705650},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\\\"],\\\"sizeBytes\\\":938303566},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:631a3798b749fecc041a99929eb946618df723e15055e805ff752a1a1273481c\\\"],\\\"sizeBytes\\\":870567329},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\\\"],\\\"sizeBytes\\\":857069957},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b12f830c3316aa4dc061c2d00c74126282b3e2bcccc301eab00d57fff3c4c7c\\\"],\\\"sizeBytes\\\":767284906},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cb3ec61f9a932a9ad13bdeb44bcf9477a8d5f728151d7f19ed3ef7d4b02b3a82\\\"],\\\"sizeBytes\\\":682371258},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:916566bb9d0143352324233d460ad94697719c11c8c9158e3aea8f475941751f\\\"],\\\"sizeBytes\\\":677523572},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b\\\"],\\\"sizeBytes\\\":672407260},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9724d2036305cbd729e1f484c5bad89971de977fff8a6723fef1873858dd1123\\\"],\\\"sizeBytes\\\":616108962},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:df606f3b71d4376d1a2108c09f0d3dab455fc30bcb67c60e91590c105e9025bf\\\"],\\\"sizeBytes\\\":583836304},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2\\\"],\\\"sizeBytes\\\":576619763},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8\\\"],\\\"sizeBytes\\\":552673986},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc\\\"],\\\"sizeBytes\\\":543227406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718\\\"],\\\"sizeBytes\\\":532719167},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cfde59e48cd5dee3721f34d249cb119cc3259fd857965d34f9c7ed83b0c363a1\\\"],\\\"sizeBytes\\\":532402162},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f4d4282cb53325e737ad68abbfcb70687ae04fb50353f4f0ba0ba5703b15009a\\\"],\\\"sizeBytes\\\":512838054},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\\\"],\\\"sizeBytes\\\":512452153},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660\\\"],\\\"sizeBytes\\\":509437356},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df\\\"],\\\"sizeBytes\\\":507687221},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8375671da86aa527ee7e291d86971b0baa823ffc7663b5a983084456e76c0f59\\\"],\\\"sizeBytes\\\":506741476},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b\\\"],\\\"sizeBytes\\\":505649178},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:86af77350cfe6fd69280157e4162aa0147873d9431c641ae4ad3e881ff768a73\\\"],\\\"sizeBytes\\\":505628211},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\\\"],\\\"sizeBytes\\\":503340749},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8139ed65c0a0a4b0f253b715c11cc52be027efe8a4774da9ccce35c78ef439da\\\"],\\\"sizeBytes\\\":503011144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4\\\"],\\\"sizeBytes\\\":502436444},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10e57ca7611f79710f05777dc6a8f31c7e04eb09da4d8d793a5acfbf0e4692d7\\\"],\\\"sizeBytes\\\":500943492},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce\\\"],\\\"sizeBytes\\\":500848684},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c\\\"],\\\"sizeBytes\\\":499798563},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96\\\"],\\\"sizeBytes\\\":499705918},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:75d996f6147edb88c09fd1a052099de66638590d7d03a735006244bc9e19f898\\\"],\\\"sizeBytes\\\":499082775},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3b8d91a25eeb9f02041e947adb3487da3e7ab8449d3d2ad015827e7954df7b34\\\"],\\\"sizeBytes\\\":490455952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68\\\"],\\\"sizeBytes\\\":489528665},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1\\\"],\\\"sizeBytes\\\":481559117},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3a77aa4d03b89ea284e3467a268e5989a77a2ef63e685eb1d5c5ea5b3922b7a\\\"],\\\"sizeBytes\\\":478917802},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eb928c13a46d3fb45f4a881892d023a92d610a5430be0ffd916aaf8da8e7d297\\\"],\\\"sizeBytes\\\":478642572},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fd3e9f8f00a59bda7483ec7dc8a0ed602f9ca30e3d72b22072dbdf2819da3f61\\\"],\\\"sizeBytes\\\":465144618},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3c1edf52f70bf9b1d1457e0c4111bc79cdaa1edd659ddbdb9d8176eff8b46956\\\"],\\\"sizeBytes\\\":462727837},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8cc27777e72233024fe84ee1faa168aec715a0b24912a3ce70715ddccba328df\\\"],\\\"sizeBytes\\\":461702648},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\\\"],\\\"sizeBytes\\\":459552216},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d3ce2cbf1032ad0f24f204db73687002fcf302e86ebde3945801c74351b64576\\\"],\\\"sizeBytes\\\":458169255},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7664a2d4cb10e82ed32abbf95799f43fc3d10135d7dd94799730de504a89680a\\\"],\\\"sizeBytes\\\":452589750},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4ecc5bac651ff1942865baee5159582e9602c89b47eeab18400a32abcba8f690\\\"],\\\"sizeBytes\\\":451039520},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2632d7f05d5a992e91038ded81c715898f3fe803420a9b67a0201e9fd8075213\\\"],\\\"sizeBytes\\\":443291941},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3f0aa9cd04713acc5c6fea721bd849e1500da8ae945e0b32000887f34d786e0b\\\"],\\\"sizeBytes\\\":442509555},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e438b814f8e16f00b3fc4b69991af80eee79ae111d2a707f34aa64b2ccbb6eb\\\"],\\\"sizeBytes\\\":437737925},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b00c658332d6c6786bd969b26097c20a78c79c045f1692a8809234f5fb586c22\\\"],\\\"sizeBytes\\\":433122306},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a3d37aa7a22c68afa963ecfb4b43c52cccf152580cd66e4d5382fb69e4037cc\\\"],\\\"sizeBytes\\\":406053031},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9432c13d76bd4ba4eb9197c050cf88c0d701fa2055eeb59257e2e23901f9fdff\\\"],\\\"sizeBytes\\\":401810450}]}}\" for node \"master-0\": Patch \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0/status?timeout=10s\": context deadline exceeded" Dec 05 10:40:11.419355 master-0 kubenswrapper[10444]: E1205 10:40:11.419276 10444 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:40:11.618667 master-0 kubenswrapper[10444]: E1205 10:40:11.618558 10444 mirror_client.go:138] "Failed deleting a mirror pod" err="Timeout: request did not complete within requested timeout - context deadline exceeded" pod="openshift-etcd/etcd-master-0-master-0" Dec 05 10:40:11.618937 master-0 kubenswrapper[10444]: E1205 10:40:11.618797 10444 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="34.012s" Dec 05 10:40:11.631077 master-0 kubenswrapper[10444]: I1205 10:40:11.630982 10444 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0-master-0" podUID="" Dec 05 10:40:11.666476 master-0 kubenswrapper[10444]: E1205 10:40:11.666366 10444 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="1.6s" Dec 05 10:40:14.834602 master-0 kubenswrapper[10444]: E1205 10:40:14.834356 10444 event.go:359] "Server rejected event (will not retry!)" err="Timeout: request did not complete within requested timeout - context deadline exceeded" event="&Event{ObjectMeta:{machine-approver-f797d8546-65t96.187e4b7f10fd4b17 openshift-cluster-machine-approver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-cluster-machine-approver,Name:machine-approver-f797d8546-65t96,UID:e4c3f514-33e8-4deb-932e-56ddf7281c4c,APIVersion:v1,ResourceVersion:8584,FieldPath:spec.containers{machine-approver-controller},},Reason:Started,Message:Started container machine-approver-controller,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:38:33.093778199 +0000 UTC m=+53.685690776,LastTimestamp:2025-12-05 10:38:33.093778199 +0000 UTC m=+53.685690776,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:40:21.419795 master-0 kubenswrapper[10444]: E1205 10:40:21.419679 10444 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:40:23.131411 master-0 kubenswrapper[10444]: I1205 10:40:23.131319 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5_1e69ce9e-4e6f-4015-9ba6-5a7942570190/openshift-controller-manager-operator/1.log" Dec 05 10:40:23.132298 master-0 kubenswrapper[10444]: I1205 10:40:23.132052 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5_1e69ce9e-4e6f-4015-9ba6-5a7942570190/openshift-controller-manager-operator/0.log" Dec 05 10:40:23.132298 master-0 kubenswrapper[10444]: I1205 10:40:23.132095 10444 generic.go:334] "Generic (PLEG): container finished" podID="1e69ce9e-4e6f-4015-9ba6-5a7942570190" containerID="ded2f19eb6b7e89481ac839fb3ba5dd47ee60cff52debe8dba17401c93d0ac1a" exitCode=255 Dec 05 10:40:23.268176 master-0 kubenswrapper[10444]: E1205 10:40:23.267974 10444 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 05 10:40:26.955330 master-0 kubenswrapper[10444]: E1205 10:40:26.955266 10444 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="15.336s" Dec 05 10:40:26.955330 master-0 kubenswrapper[10444]: I1205 10:40:26.955325 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-1-master-0" event={"ID":"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9","Type":"ContainerDied","Data":"0d4c12890814d7bd72e2d33d03a6131a3d52c0e2122d3701215dc96f26a231b2"} Dec 05 10:40:26.956152 master-0 kubenswrapper[10444]: I1205 10:40:26.955360 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-1-master-0" event={"ID":"738d9fe7-2d08-4aa4-842b-9cbeddf120fb","Type":"ContainerDied","Data":"8a8e52c1bfb7c50d31dbfbb099e72a94f8cf7f5c9eac30d42a9e7bb6f30721bf"} Dec 05 10:40:26.958374 master-0 kubenswrapper[10444]: I1205 10:40:26.958317 10444 scope.go:117] "RemoveContainer" containerID="f2785fc5d00b92168028cf4e3273a28c6085e8db64c579bdce5eca572512c7d9" Dec 05 10:40:26.958545 master-0 kubenswrapper[10444]: I1205 10:40:26.958501 10444 scope.go:117] "RemoveContainer" containerID="7f719d929cd8f822ace129247877e00bae5a9457d73d9da2566b48784c2ee8fc" Dec 05 10:40:26.963725 master-0 kubenswrapper[10444]: I1205 10:40:26.963173 10444 scope.go:117] "RemoveContainer" containerID="f1e6d80e2ae7d337901560dde2050e55f54f6b182ab45a5e1cce9ebc5f4043a9" Dec 05 10:40:26.964149 master-0 kubenswrapper[10444]: I1205 10:40:26.964130 10444 scope.go:117] "RemoveContainer" containerID="ded2f19eb6b7e89481ac839fb3ba5dd47ee60cff52debe8dba17401c93d0ac1a" Dec 05 10:40:26.964412 master-0 kubenswrapper[10444]: E1205 10:40:26.964385 10444 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openshift-controller-manager-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=openshift-controller-manager-operator pod=openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator(1e69ce9e-4e6f-4015-9ba6-5a7942570190)\"" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" podUID="1e69ce9e-4e6f-4015-9ba6-5a7942570190" Dec 05 10:40:26.964635 master-0 kubenswrapper[10444]: I1205 10:40:26.964608 10444 scope.go:117] "RemoveContainer" containerID="a4e599b9d9dac0eea61bdc89656e6a7ee18f83aa3fd9ba9340f57da02bd4a39b" Dec 05 10:40:26.964695 master-0 kubenswrapper[10444]: I1205 10:40:26.964642 10444 scope.go:117] "RemoveContainer" containerID="e7ce8dc7bd304b25a361ecabc9d8faa1f0ae1eca9ef1f9dc24bcebc4d09b4a20" Dec 05 10:40:26.965079 master-0 kubenswrapper[10444]: I1205 10:40:26.965005 10444 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0-master-0" podUID="" Dec 05 10:40:26.965079 master-0 kubenswrapper[10444]: I1205 10:40:26.965057 10444 scope.go:117] "RemoveContainer" containerID="5674d03e599731519f29f243f076ae28159025706182191e04e4bbfef0819512" Dec 05 10:40:26.965168 master-0 kubenswrapper[10444]: I1205 10:40:26.965149 10444 scope.go:117] "RemoveContainer" containerID="cc8433167039cff15f2a8553e5790127434ae4c5286a5db0777d417011f80f8c" Dec 05 10:40:26.996379 master-0 kubenswrapper[10444]: I1205 10:40:26.995869 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:40:26.996379 master-0 kubenswrapper[10444]: I1205 10:40:26.995922 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerDied","Data":"41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2"} Dec 05 10:40:26.996379 master-0 kubenswrapper[10444]: I1205 10:40:26.996003 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-3-master-0" event={"ID":"6325f9a2-86d5-4fac-9e71-331b2628f965","Type":"ContainerDied","Data":"df5bdcc899d9a7c7268a300825923259315bb7a02282ed76b8b7644641b06995"} Dec 05 10:40:26.996379 master-0 kubenswrapper[10444]: I1205 10:40:26.996024 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-ql7j7" event={"ID":"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e","Type":"ContainerDied","Data":"5674d03e599731519f29f243f076ae28159025706182191e04e4bbfef0819512"} Dec 05 10:40:26.996379 master-0 kubenswrapper[10444]: I1205 10:40:26.996039 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" event={"ID":"4825316a-ea9f-4d3d-838b-fa809a6e49c7","Type":"ContainerDied","Data":"f2785fc5d00b92168028cf4e3273a28c6085e8db64c579bdce5eca572512c7d9"} Dec 05 10:40:26.996379 master-0 kubenswrapper[10444]: I1205 10:40:26.996054 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" event={"ID":"f7a08359-0379-4364-8b0c-ddb58ff605f4","Type":"ContainerDied","Data":"7f719d929cd8f822ace129247877e00bae5a9457d73d9da2566b48784c2ee8fc"} Dec 05 10:40:26.996379 master-0 kubenswrapper[10444]: I1205 10:40:26.996068 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5"} Dec 05 10:40:26.996379 master-0 kubenswrapper[10444]: I1205 10:40:26.996130 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e"} Dec 05 10:40:26.996379 master-0 kubenswrapper[10444]: I1205 10:40:26.996146 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab"} Dec 05 10:40:26.996379 master-0 kubenswrapper[10444]: I1205 10:40:26.996159 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819"} Dec 05 10:40:26.996965 master-0 kubenswrapper[10444]: I1205 10:40:26.996704 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526"} Dec 05 10:40:26.996965 master-0 kubenswrapper[10444]: I1205 10:40:26.996727 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" event={"ID":"6f76d12f-5406-47e2-8337-2f50e35376d6","Type":"ContainerDied","Data":"cc8433167039cff15f2a8553e5790127434ae4c5286a5db0777d417011f80f8c"} Dec 05 10:40:26.996965 master-0 kubenswrapper[10444]: I1205 10:40:26.996741 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" event={"ID":"1a0235af-2cf2-4ad4-b419-764fb56a0107","Type":"ContainerDied","Data":"f1e6d80e2ae7d337901560dde2050e55f54f6b182ab45a5e1cce9ebc5f4043a9"} Dec 05 10:40:26.996965 master-0 kubenswrapper[10444]: I1205 10:40:26.996754 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" event={"ID":"444f8808-e454-4015-9e20-429e715a08c7","Type":"ContainerDied","Data":"a4e599b9d9dac0eea61bdc89656e6a7ee18f83aa3fd9ba9340f57da02bd4a39b"} Dec 05 10:40:26.996965 master-0 kubenswrapper[10444]: I1205 10:40:26.996768 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" event={"ID":"d95a56ba-c940-4e3e-aed6-d8c04f1871b6","Type":"ContainerDied","Data":"dad5a560aae0ea7cbefd19607cf4263a0d5cfc63a06ebe4360483bee7047a720"} Dec 05 10:40:26.996965 master-0 kubenswrapper[10444]: I1205 10:40:26.996783 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" event={"ID":"d95a56ba-c940-4e3e-aed6-d8c04f1871b6","Type":"ContainerStarted","Data":"61e3c1e2bb3e972291fccf7fa16658ac2da0a3dc54e79b6f27f8a488bee5a098"} Dec 05 10:40:26.996965 master-0 kubenswrapper[10444]: I1205 10:40:26.996799 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerDied","Data":"e7ce8dc7bd304b25a361ecabc9d8faa1f0ae1eca9ef1f9dc24bcebc4d09b4a20"} Dec 05 10:40:26.996965 master-0 kubenswrapper[10444]: I1205 10:40:26.996815 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" event={"ID":"1e69ce9e-4e6f-4015-9ba6-5a7942570190","Type":"ContainerDied","Data":"ded2f19eb6b7e89481ac839fb3ba5dd47ee60cff52debe8dba17401c93d0ac1a"} Dec 05 10:40:26.997590 master-0 kubenswrapper[10444]: I1205 10:40:26.997311 10444 scope.go:117] "RemoveContainer" containerID="303d7cff916f69b65cb8e8b72a47e944bb182f86fc3e9bfde505fa8e541eb3b6" Dec 05 10:40:26.999193 master-0 kubenswrapper[10444]: I1205 10:40:26.999140 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-etcd/etcd-master-0-master-0"] Dec 05 10:40:26.999259 master-0 kubenswrapper[10444]: I1205 10:40:26.999192 10444 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-etcd/etcd-master-0-master-0" mirrorPodUID="0db71427-81ac-457d-b241-4f2ee999cfad" Dec 05 10:40:27.000960 master-0 kubenswrapper[10444]: I1205 10:40:27.000944 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-etcd/etcd-master-0-master-0"] Dec 05 10:40:27.002144 master-0 kubenswrapper[10444]: I1205 10:40:27.002106 10444 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-etcd/etcd-master-0-master-0" mirrorPodUID="0db71427-81ac-457d-b241-4f2ee999cfad" Dec 05 10:40:27.066215 master-0 kubenswrapper[10444]: I1205 10:40:27.066108 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" podStartSLOduration=115.587568721 podStartE2EDuration="1m57.066089461s" podCreationTimestamp="2025-12-05 10:38:30 +0000 UTC" firstStartedPulling="2025-12-05 10:38:31.347490079 +0000 UTC m=+51.939402636" lastFinishedPulling="2025-12-05 10:38:32.826010809 +0000 UTC m=+53.417923376" observedRunningTime="2025-12-05 10:40:27.064368887 +0000 UTC m=+167.656281454" watchObservedRunningTime="2025-12-05 10:40:27.066089461 +0000 UTC m=+167.658002028" Dec 05 10:40:27.110136 master-0 kubenswrapper[10444]: I1205 10:40:27.110096 10444 scope.go:117] "RemoveContainer" containerID="41f047cf59b8b167740e4f4b07cc948566d4077a26e6c889e254fc22e5a5bd11" Dec 05 10:40:27.160078 master-0 kubenswrapper[10444]: I1205 10:40:27.160052 10444 scope.go:117] "RemoveContainer" containerID="ded2f19eb6b7e89481ac839fb3ba5dd47ee60cff52debe8dba17401c93d0ac1a" Dec 05 10:40:27.160567 master-0 kubenswrapper[10444]: E1205 10:40:27.160522 10444 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openshift-controller-manager-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=openshift-controller-manager-operator pod=openshift-controller-manager-operator-6c8676f99d-cwvk5_openshift-controller-manager-operator(1e69ce9e-4e6f-4015-9ba6-5a7942570190)\"" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" podUID="1e69ce9e-4e6f-4015-9ba6-5a7942570190" Dec 05 10:40:27.564534 master-0 kubenswrapper[10444]: I1205 10:40:27.564343 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_installer-1-master-0_738d9fe7-2d08-4aa4-842b-9cbeddf120fb/installer/0.log" Dec 05 10:40:27.564534 master-0 kubenswrapper[10444]: I1205 10:40:27.564403 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:40:27.611119 master-0 kubenswrapper[10444]: I1205 10:40:27.611077 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kubelet-dir\") pod \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " Dec 05 10:40:27.611190 master-0 kubenswrapper[10444]: I1205 10:40:27.611161 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-var-lock\") pod \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " Dec 05 10:40:27.613466 master-0 kubenswrapper[10444]: I1205 10:40:27.611205 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kube-api-access\") pod \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\" (UID: \"738d9fe7-2d08-4aa4-842b-9cbeddf120fb\") " Dec 05 10:40:27.613466 master-0 kubenswrapper[10444]: I1205 10:40:27.611271 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "738d9fe7-2d08-4aa4-842b-9cbeddf120fb" (UID: "738d9fe7-2d08-4aa4-842b-9cbeddf120fb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:40:27.613466 master-0 kubenswrapper[10444]: I1205 10:40:27.611282 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-var-lock" (OuterVolumeSpecName: "var-lock") pod "738d9fe7-2d08-4aa4-842b-9cbeddf120fb" (UID: "738d9fe7-2d08-4aa4-842b-9cbeddf120fb"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:40:27.613466 master-0 kubenswrapper[10444]: I1205 10:40:27.611443 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:27.613466 master-0 kubenswrapper[10444]: I1205 10:40:27.611464 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:27.620431 master-0 kubenswrapper[10444]: I1205 10:40:27.614410 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "738d9fe7-2d08-4aa4-842b-9cbeddf120fb" (UID: "738d9fe7-2d08-4aa4-842b-9cbeddf120fb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:40:27.621899 master-0 kubenswrapper[10444]: I1205 10:40:27.621623 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-1-master-0_5d8c854f-31f8-42d3-b28c-82c7d1b14ee9/installer/0.log" Dec 05 10:40:27.621899 master-0 kubenswrapper[10444]: I1205 10:40:27.621688 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:40:27.627475 master-0 kubenswrapper[10444]: I1205 10:40:27.624842 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler_installer-3-master-0_6325f9a2-86d5-4fac-9e71-331b2628f965/installer/0.log" Dec 05 10:40:27.627475 master-0 kubenswrapper[10444]: I1205 10:40:27.624911 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:40:27.712409 master-0 kubenswrapper[10444]: I1205 10:40:27.712347 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-kubelet-dir\") pod \"6325f9a2-86d5-4fac-9e71-331b2628f965\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " Dec 05 10:40:27.712409 master-0 kubenswrapper[10444]: I1205 10:40:27.712401 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kubelet-dir\") pod \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " Dec 05 10:40:27.712683 master-0 kubenswrapper[10444]: I1205 10:40:27.712452 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-var-lock\") pod \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " Dec 05 10:40:27.712683 master-0 kubenswrapper[10444]: I1205 10:40:27.712484 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kube-api-access\") pod \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\" (UID: \"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9\") " Dec 05 10:40:27.712683 master-0 kubenswrapper[10444]: I1205 10:40:27.712520 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6325f9a2-86d5-4fac-9e71-331b2628f965-kube-api-access\") pod \"6325f9a2-86d5-4fac-9e71-331b2628f965\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " Dec 05 10:40:27.712683 master-0 kubenswrapper[10444]: I1205 10:40:27.712550 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-var-lock\") pod \"6325f9a2-86d5-4fac-9e71-331b2628f965\" (UID: \"6325f9a2-86d5-4fac-9e71-331b2628f965\") " Dec 05 10:40:27.712818 master-0 kubenswrapper[10444]: I1205 10:40:27.712796 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/738d9fe7-2d08-4aa4-842b-9cbeddf120fb-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:27.714163 master-0 kubenswrapper[10444]: I1205 10:40:27.712866 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-var-lock" (OuterVolumeSpecName: "var-lock") pod "6325f9a2-86d5-4fac-9e71-331b2628f965" (UID: "6325f9a2-86d5-4fac-9e71-331b2628f965"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:40:27.714163 master-0 kubenswrapper[10444]: I1205 10:40:27.712916 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6325f9a2-86d5-4fac-9e71-331b2628f965" (UID: "6325f9a2-86d5-4fac-9e71-331b2628f965"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:40:27.714163 master-0 kubenswrapper[10444]: I1205 10:40:27.712941 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5d8c854f-31f8-42d3-b28c-82c7d1b14ee9" (UID: "5d8c854f-31f8-42d3-b28c-82c7d1b14ee9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:40:27.714163 master-0 kubenswrapper[10444]: I1205 10:40:27.712960 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-var-lock" (OuterVolumeSpecName: "var-lock") pod "5d8c854f-31f8-42d3-b28c-82c7d1b14ee9" (UID: "5d8c854f-31f8-42d3-b28c-82c7d1b14ee9"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:40:27.717085 master-0 kubenswrapper[10444]: I1205 10:40:27.717017 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6325f9a2-86d5-4fac-9e71-331b2628f965-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6325f9a2-86d5-4fac-9e71-331b2628f965" (UID: "6325f9a2-86d5-4fac-9e71-331b2628f965"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:40:27.718033 master-0 kubenswrapper[10444]: I1205 10:40:27.717980 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5d8c854f-31f8-42d3-b28c-82c7d1b14ee9" (UID: "5d8c854f-31f8-42d3-b28c-82c7d1b14ee9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:40:27.814239 master-0 kubenswrapper[10444]: I1205 10:40:27.814126 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:27.814239 master-0 kubenswrapper[10444]: I1205 10:40:27.814159 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6325f9a2-86d5-4fac-9e71-331b2628f965-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:27.814239 master-0 kubenswrapper[10444]: I1205 10:40:27.814168 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:27.814239 master-0 kubenswrapper[10444]: I1205 10:40:27.814176 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:27.814239 master-0 kubenswrapper[10444]: I1205 10:40:27.814184 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5d8c854f-31f8-42d3-b28c-82c7d1b14ee9-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:27.814239 master-0 kubenswrapper[10444]: I1205 10:40:27.814192 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6325f9a2-86d5-4fac-9e71-331b2628f965-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:28.167043 master-0 kubenswrapper[10444]: I1205 10:40:28.166825 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" event={"ID":"8b47694fcc32464ab24d09c23d6efb57","Type":"ContainerStarted","Data":"efee17240223c7e06e121d89cbef5a60b4ccceeebdd17b05f12e6a98b82892f6"} Dec 05 10:40:28.169937 master-0 kubenswrapper[10444]: I1205 10:40:28.169828 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" event={"ID":"444f8808-e454-4015-9e20-429e715a08c7","Type":"ContainerStarted","Data":"081be958f469d91b44369f7b9a2fd1d6880cb567d726cc740245fe45c1fec94e"} Dec 05 10:40:28.172484 master-0 kubenswrapper[10444]: I1205 10:40:28.172455 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/0.log" Dec 05 10:40:28.172573 master-0 kubenswrapper[10444]: I1205 10:40:28.172513 10444 generic.go:334] "Generic (PLEG): container finished" podID="22676fac-b770-4937-9bee-7478bd1babb7" containerID="7f5edea6bac7d86c19774c19407fa60083727d4b313c43d3272af6d979af6367" exitCode=1 Dec 05 10:40:28.172627 master-0 kubenswrapper[10444]: I1205 10:40:28.172601 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" event={"ID":"22676fac-b770-4937-9bee-7478bd1babb7","Type":"ContainerDied","Data":"7f5edea6bac7d86c19774c19407fa60083727d4b313c43d3272af6d979af6367"} Dec 05 10:40:28.176513 master-0 kubenswrapper[10444]: I1205 10:40:28.173645 10444 scope.go:117] "RemoveContainer" containerID="7f5edea6bac7d86c19774c19407fa60083727d4b313c43d3272af6d979af6367" Dec 05 10:40:28.176513 master-0 kubenswrapper[10444]: I1205 10:40:28.174820 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" event={"ID":"4825316a-ea9f-4d3d-838b-fa809a6e49c7","Type":"ContainerStarted","Data":"66fcfd4e1afdbd3398cb750ace935d91ba542db71f3971aeea67f31628c3ef00"} Dec 05 10:40:28.177095 master-0 kubenswrapper[10444]: I1205 10:40:28.176991 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" event={"ID":"f7a08359-0379-4364-8b0c-ddb58ff605f4","Type":"ContainerStarted","Data":"4ebfc00b497d18970f0f221db9978daf3b3a56e51ef76dfd9c99c73031c71b97"} Dec 05 10:40:28.185001 master-0 kubenswrapper[10444]: I1205 10:40:28.184029 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-node-identity_network-node-identity-ql7j7_d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/approver/0.log" Dec 05 10:40:28.185001 master-0 kubenswrapper[10444]: I1205 10:40:28.184682 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-ql7j7" event={"ID":"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e","Type":"ContainerStarted","Data":"75e16cb5a0af404fec070efd073424497756db4fbc0649e23d5c8b08a868d67f"} Dec 05 10:40:28.187236 master-0 kubenswrapper[10444]: I1205 10:40:28.187209 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler_installer-3-master-0_6325f9a2-86d5-4fac-9e71-331b2628f965/installer/0.log" Dec 05 10:40:28.187563 master-0 kubenswrapper[10444]: I1205 10:40:28.187312 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-3-master-0" event={"ID":"6325f9a2-86d5-4fac-9e71-331b2628f965","Type":"ContainerDied","Data":"ba7c2d250edf116aab8b9c6bc4e16427c460bc08a2f3f72ce479a134eaf54756"} Dec 05 10:40:28.187563 master-0 kubenswrapper[10444]: I1205 10:40:28.187355 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba7c2d250edf116aab8b9c6bc4e16427c460bc08a2f3f72ce479a134eaf54756" Dec 05 10:40:28.187563 master-0 kubenswrapper[10444]: I1205 10:40:28.187390 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:40:28.189370 master-0 kubenswrapper[10444]: I1205 10:40:28.189334 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-1-master-0_5d8c854f-31f8-42d3-b28c-82c7d1b14ee9/installer/0.log" Dec 05 10:40:28.189613 master-0 kubenswrapper[10444]: I1205 10:40:28.189556 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-1-master-0" event={"ID":"5d8c854f-31f8-42d3-b28c-82c7d1b14ee9","Type":"ContainerDied","Data":"0127ee22dc31281cc8a51e81727d0320a63ff8209b1fadb0e73c9f87ae0d74aa"} Dec 05 10:40:28.189661 master-0 kubenswrapper[10444]: I1205 10:40:28.189620 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:40:28.189827 master-0 kubenswrapper[10444]: I1205 10:40:28.189633 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0127ee22dc31281cc8a51e81727d0320a63ff8209b1fadb0e73c9f87ae0d74aa" Dec 05 10:40:28.192218 master-0 kubenswrapper[10444]: I1205 10:40:28.192195 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_network-operator-79767b7ff9-t8j2j_1a0235af-2cf2-4ad4-b419-764fb56a0107/network-operator/0.log" Dec 05 10:40:28.192318 master-0 kubenswrapper[10444]: I1205 10:40:28.192288 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" event={"ID":"1a0235af-2cf2-4ad4-b419-764fb56a0107","Type":"ContainerStarted","Data":"a08f843706833dfc728e88e2c2cea79891e8c8393ef391e12533de6bef924933"} Dec 05 10:40:28.194823 master-0 kubenswrapper[10444]: I1205 10:40:28.194772 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" event={"ID":"6f76d12f-5406-47e2-8337-2f50e35376d6","Type":"ContainerStarted","Data":"a9c7e64b14cb47a9ab52158339c38bf528cc5bfc614193152b3c1935a66a8941"} Dec 05 10:40:28.196926 master-0 kubenswrapper[10444]: I1205 10:40:28.196898 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-controller_operator-controller-controller-manager-7cbd59c7f8-dh5tt_592373ae-a952-4dd3-a9bc-f9c9c19c0802/manager/0.log" Dec 05 10:40:28.197000 master-0 kubenswrapper[10444]: I1205 10:40:28.196944 10444 generic.go:334] "Generic (PLEG): container finished" podID="592373ae-a952-4dd3-a9bc-f9c9c19c0802" containerID="665b41c4b05327dcf1d9daca6aa21544857362b43e6c7d8ef7fac95d7153f9b2" exitCode=1 Dec 05 10:40:28.197042 master-0 kubenswrapper[10444]: I1205 10:40:28.197007 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" event={"ID":"592373ae-a952-4dd3-a9bc-f9c9c19c0802","Type":"ContainerDied","Data":"665b41c4b05327dcf1d9daca6aa21544857362b43e6c7d8ef7fac95d7153f9b2"} Dec 05 10:40:28.197667 master-0 kubenswrapper[10444]: I1205 10:40:28.197642 10444 scope.go:117] "RemoveContainer" containerID="665b41c4b05327dcf1d9daca6aa21544857362b43e6c7d8ef7fac95d7153f9b2" Dec 05 10:40:28.209824 master-0 kubenswrapper[10444]: I1205 10:40:28.206717 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_installer-1-master-0_738d9fe7-2d08-4aa4-842b-9cbeddf120fb/installer/0.log" Dec 05 10:40:28.209824 master-0 kubenswrapper[10444]: I1205 10:40:28.206965 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:40:28.209824 master-0 kubenswrapper[10444]: I1205 10:40:28.207151 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-1-master-0" event={"ID":"738d9fe7-2d08-4aa4-842b-9cbeddf120fb","Type":"ContainerDied","Data":"481bec329814be919f1e3dd3cd8fba26d8d823339446a0f9b66a76525352b23f"} Dec 05 10:40:28.209824 master-0 kubenswrapper[10444]: I1205 10:40:28.207261 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="481bec329814be919f1e3dd3cd8fba26d8d823339446a0f9b66a76525352b23f" Dec 05 10:40:28.220107 master-0 kubenswrapper[10444]: I1205 10:40:28.220024 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5_1e69ce9e-4e6f-4015-9ba6-5a7942570190/openshift-controller-manager-operator/1.log" Dec 05 10:40:28.897132 master-0 kubenswrapper[10444]: I1205 10:40:28.897077 10444 patch_prober.go:28] interesting pod/catalogd-controller-manager-7cc89f4c4c-lth87 container/manager namespace/openshift-catalogd: Readiness probe status=failure output="Get \"http://10.128.0.40:8081/readyz\": dial tcp 10.128.0.40:8081: connect: connection refused" start-of-body= Dec 05 10:40:28.897255 master-0 kubenswrapper[10444]: I1205 10:40:28.897138 10444 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" podUID="a4fd453c-e667-4bdc-aa9e-3d95ff707200" containerName="manager" probeResult="failure" output="Get \"http://10.128.0.40:8081/readyz\": dial tcp 10.128.0.40:8081: connect: connection refused" Dec 05 10:40:28.897255 master-0 kubenswrapper[10444]: I1205 10:40:28.897086 10444 patch_prober.go:28] interesting pod/catalogd-controller-manager-7cc89f4c4c-lth87 container/manager namespace/openshift-catalogd: Liveness probe status=failure output="Get \"http://10.128.0.40:8081/healthz\": dial tcp 10.128.0.40:8081: connect: connection refused" start-of-body= Dec 05 10:40:28.897352 master-0 kubenswrapper[10444]: I1205 10:40:28.897236 10444 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" podUID="a4fd453c-e667-4bdc-aa9e-3d95ff707200" containerName="manager" probeResult="failure" output="Get \"http://10.128.0.40:8081/healthz\": dial tcp 10.128.0.40:8081: connect: connection refused" Dec 05 10:40:29.227329 master-0 kubenswrapper[10444]: I1205 10:40:29.227178 10444 generic.go:334] "Generic (PLEG): container finished" podID="fd58232c-a81a-4aee-8b2c-5ffcdded2e23" containerID="3908741f62d9038e3367c49d1d01f504f11849e5b644100de90c2c7d751cc2e3" exitCode=0 Dec 05 10:40:29.227913 master-0 kubenswrapper[10444]: I1205 10:40:29.227271 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" event={"ID":"fd58232c-a81a-4aee-8b2c-5ffcdded2e23","Type":"ContainerDied","Data":"3908741f62d9038e3367c49d1d01f504f11849e5b644100de90c2c7d751cc2e3"} Dec 05 10:40:29.228684 master-0 kubenswrapper[10444]: I1205 10:40:29.228648 10444 generic.go:334] "Generic (PLEG): container finished" podID="926263c4-ec5b-41cb-9c30-0c88f636035f" containerID="3504f06691ba2b115136a2091ae95118383b7c2cc711e4550b1e60138be21510" exitCode=0 Dec 05 10:40:29.228752 master-0 kubenswrapper[10444]: I1205 10:40:29.228654 10444 scope.go:117] "RemoveContainer" containerID="3908741f62d9038e3367c49d1d01f504f11849e5b644100de90c2c7d751cc2e3" Dec 05 10:40:29.228752 master-0 kubenswrapper[10444]: I1205 10:40:29.228717 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" event={"ID":"926263c4-ec5b-41cb-9c30-0c88f636035f","Type":"ContainerDied","Data":"3504f06691ba2b115136a2091ae95118383b7c2cc711e4550b1e60138be21510"} Dec 05 10:40:29.229101 master-0 kubenswrapper[10444]: I1205 10:40:29.229070 10444 scope.go:117] "RemoveContainer" containerID="3504f06691ba2b115136a2091ae95118383b7c2cc711e4550b1e60138be21510" Dec 05 10:40:29.230708 master-0 kubenswrapper[10444]: I1205 10:40:29.230610 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-catalogd_catalogd-controller-manager-7cc89f4c4c-lth87_a4fd453c-e667-4bdc-aa9e-3d95ff707200/manager/0.log" Dec 05 10:40:29.231609 master-0 kubenswrapper[10444]: I1205 10:40:29.231122 10444 generic.go:334] "Generic (PLEG): container finished" podID="a4fd453c-e667-4bdc-aa9e-3d95ff707200" containerID="2e9155fd74fd1d13dcef2794db107599bb6fa96cc533e3303f4bbbfdbd8c0208" exitCode=1 Dec 05 10:40:29.231609 master-0 kubenswrapper[10444]: I1205 10:40:29.231176 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" event={"ID":"a4fd453c-e667-4bdc-aa9e-3d95ff707200","Type":"ContainerDied","Data":"2e9155fd74fd1d13dcef2794db107599bb6fa96cc533e3303f4bbbfdbd8c0208"} Dec 05 10:40:29.231609 master-0 kubenswrapper[10444]: I1205 10:40:29.231529 10444 scope.go:117] "RemoveContainer" containerID="2e9155fd74fd1d13dcef2794db107599bb6fa96cc533e3303f4bbbfdbd8c0208" Dec 05 10:40:29.235901 master-0 kubenswrapper[10444]: I1205 10:40:29.235857 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-controller_operator-controller-controller-manager-7cbd59c7f8-dh5tt_592373ae-a952-4dd3-a9bc-f9c9c19c0802/manager/0.log" Dec 05 10:40:29.235991 master-0 kubenswrapper[10444]: I1205 10:40:29.235937 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" event={"ID":"592373ae-a952-4dd3-a9bc-f9c9c19c0802","Type":"ContainerStarted","Data":"405cb7fe7c4f15258168a2b739de72c941100f0f87222e1596be37c3a536b301"} Dec 05 10:40:29.236183 master-0 kubenswrapper[10444]: I1205 10:40:29.236143 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:40:29.237343 master-0 kubenswrapper[10444]: I1205 10:40:29.237296 10444 generic.go:334] "Generic (PLEG): container finished" podID="11f563d5-89bb-433c-956a-6d5d2492e8f1" containerID="475c118c4e82c9f8f86c3034eb04eddc9ae813cf612425cd157e214036c29617" exitCode=0 Dec 05 10:40:29.237439 master-0 kubenswrapper[10444]: I1205 10:40:29.237363 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" event={"ID":"11f563d5-89bb-433c-956a-6d5d2492e8f1","Type":"ContainerDied","Data":"475c118c4e82c9f8f86c3034eb04eddc9ae813cf612425cd157e214036c29617"} Dec 05 10:40:29.237705 master-0 kubenswrapper[10444]: I1205 10:40:29.237666 10444 scope.go:117] "RemoveContainer" containerID="475c118c4e82c9f8f86c3034eb04eddc9ae813cf612425cd157e214036c29617" Dec 05 10:40:29.243195 master-0 kubenswrapper[10444]: I1205 10:40:29.241826 10444 generic.go:334] "Generic (PLEG): container finished" podID="b760849c-8d83-47da-8677-68445c143bef" containerID="d5454b7ce1bc247671ab46448edd5a6f0a198a1673e48f822fea3525f6db868a" exitCode=0 Dec 05 10:40:29.243195 master-0 kubenswrapper[10444]: I1205 10:40:29.241899 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" event={"ID":"b760849c-8d83-47da-8677-68445c143bef","Type":"ContainerDied","Data":"d5454b7ce1bc247671ab46448edd5a6f0a198a1673e48f822fea3525f6db868a"} Dec 05 10:40:29.243195 master-0 kubenswrapper[10444]: I1205 10:40:29.242308 10444 scope.go:117] "RemoveContainer" containerID="d5454b7ce1bc247671ab46448edd5a6f0a198a1673e48f822fea3525f6db868a" Dec 05 10:40:29.248240 master-0 kubenswrapper[10444]: I1205 10:40:29.248215 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/0.log" Dec 05 10:40:29.248341 master-0 kubenswrapper[10444]: I1205 10:40:29.248315 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" event={"ID":"22676fac-b770-4937-9bee-7478bd1babb7","Type":"ContainerStarted","Data":"01ddc03289ad66f2a901b27cad793faf8cd47f3f421b8573890f1eb86f36b80a"} Dec 05 10:40:30.264279 master-0 kubenswrapper[10444]: I1205 10:40:30.264137 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" event={"ID":"11f563d5-89bb-433c-956a-6d5d2492e8f1","Type":"ContainerStarted","Data":"5d29cfb25e3c664f42b949f72a0cc8a81a2fbc483c367ef78fa4ab375de3d64d"} Dec 05 10:40:30.268750 master-0 kubenswrapper[10444]: I1205 10:40:30.268646 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" event={"ID":"fd58232c-a81a-4aee-8b2c-5ffcdded2e23","Type":"ContainerStarted","Data":"5afcf4c581b0a6829a04c849a064f642044e0550587c2d0f0b4eace491752b5e"} Dec 05 10:40:30.275754 master-0 kubenswrapper[10444]: I1205 10:40:30.275687 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" event={"ID":"926263c4-ec5b-41cb-9c30-0c88f636035f","Type":"ContainerStarted","Data":"9426824799d7ff29210670da649e79b8fec18b4165425c4e34958dec644e0eb2"} Dec 05 10:40:30.276123 master-0 kubenswrapper[10444]: I1205 10:40:30.276069 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:40:30.278514 master-0 kubenswrapper[10444]: I1205 10:40:30.278400 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:40:30.279717 master-0 kubenswrapper[10444]: I1205 10:40:30.279648 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-catalogd_catalogd-controller-manager-7cc89f4c4c-lth87_a4fd453c-e667-4bdc-aa9e-3d95ff707200/manager/0.log" Dec 05 10:40:30.280411 master-0 kubenswrapper[10444]: I1205 10:40:30.280298 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" event={"ID":"a4fd453c-e667-4bdc-aa9e-3d95ff707200","Type":"ContainerStarted","Data":"eaa0a2bfed5203d2ab62ebaff09a1a3afca45eb5d5bdd915133bd4c75546ae0e"} Dec 05 10:40:30.280976 master-0 kubenswrapper[10444]: I1205 10:40:30.280918 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:40:30.281247 master-0 kubenswrapper[10444]: I1205 10:40:30.281177 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-master-0" Dec 05 10:40:30.284020 master-0 kubenswrapper[10444]: I1205 10:40:30.283957 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" event={"ID":"b760849c-8d83-47da-8677-68445c143bef","Type":"ContainerStarted","Data":"0423f72dfd535a3ee285bcc8d731216360e562171918005d692bba035e2bd6b5"} Dec 05 10:40:30.680528 master-0 kubenswrapper[10444]: I1205 10:40:30.680454 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-master-0"] Dec 05 10:40:30.727157 master-0 kubenswrapper[10444]: I1205 10:40:30.727072 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:40:31.290562 master-0 kubenswrapper[10444]: I1205 10:40:31.290483 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/0.log" Dec 05 10:40:31.290562 master-0 kubenswrapper[10444]: I1205 10:40:31.290536 10444 generic.go:334] "Generic (PLEG): container finished" podID="e27c0798-ec1c-43cd-b81b-f77f2f11ad0f" containerID="4308a58237214eca1f5ba917e53d41b87b4e41bf19760a2305217ef9ccb8ed3f" exitCode=1 Dec 05 10:40:31.291602 master-0 kubenswrapper[10444]: I1205 10:40:31.291071 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" event={"ID":"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f","Type":"ContainerDied","Data":"4308a58237214eca1f5ba917e53d41b87b4e41bf19760a2305217ef9ccb8ed3f"} Dec 05 10:40:31.291602 master-0 kubenswrapper[10444]: I1205 10:40:31.291287 10444 scope.go:117] "RemoveContainer" containerID="4308a58237214eca1f5ba917e53d41b87b4e41bf19760a2305217ef9ccb8ed3f" Dec 05 10:40:31.305317 master-0 kubenswrapper[10444]: E1205 10:40:31.305255 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-master-0\" already exists" pod="openshift-etcd/etcd-master-0" Dec 05 10:40:31.323299 master-0 kubenswrapper[10444]: I1205 10:40:31.323199 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-master-0" podStartSLOduration=1.323173555 podStartE2EDuration="1.323173555s" podCreationTimestamp="2025-12-05 10:40:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:40:31.323119444 +0000 UTC m=+171.915032011" watchObservedRunningTime="2025-12-05 10:40:31.323173555 +0000 UTC m=+171.915086152" Dec 05 10:40:31.420103 master-0 kubenswrapper[10444]: E1205 10:40:31.420024 10444 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:40:32.298192 master-0 kubenswrapper[10444]: I1205 10:40:32.298072 10444 generic.go:334] "Generic (PLEG): container finished" podID="d1c3b7dd-f25e-4983-8a94-084f863fd5b9" containerID="e562ce0da1077d81598e22070b676b5642a7e1f7ad3053d9875573b9c4c32d50" exitCode=0 Dec 05 10:40:32.298192 master-0 kubenswrapper[10444]: I1205 10:40:32.298129 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" event={"ID":"d1c3b7dd-f25e-4983-8a94-084f863fd5b9","Type":"ContainerDied","Data":"e562ce0da1077d81598e22070b676b5642a7e1f7ad3053d9875573b9c4c32d50"} Dec 05 10:40:32.298807 master-0 kubenswrapper[10444]: I1205 10:40:32.298654 10444 scope.go:117] "RemoveContainer" containerID="e562ce0da1077d81598e22070b676b5642a7e1f7ad3053d9875573b9c4c32d50" Dec 05 10:40:32.304899 master-0 kubenswrapper[10444]: I1205 10:40:32.304825 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/0.log" Dec 05 10:40:32.304975 master-0 kubenswrapper[10444]: I1205 10:40:32.304935 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" event={"ID":"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f","Type":"ContainerStarted","Data":"c81ba72a2f308c68cd7463f96c3158dfa0931cd098959c3eb1d285e9ee7a63b6"} Dec 05 10:40:33.312700 master-0 kubenswrapper[10444]: I1205 10:40:33.312626 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" event={"ID":"d1c3b7dd-f25e-4983-8a94-084f863fd5b9","Type":"ContainerStarted","Data":"b8308133ae69a2b02f4b5f12b9d8936b1d2deba6ca023e16042d1c7606ad61c0"} Dec 05 10:40:35.282128 master-0 kubenswrapper[10444]: I1205 10:40:35.282046 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-master-0" Dec 05 10:40:35.316778 master-0 kubenswrapper[10444]: I1205 10:40:35.316676 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-master-0" Dec 05 10:40:35.338769 master-0 kubenswrapper[10444]: I1205 10:40:35.338686 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-master-0" Dec 05 10:40:38.067136 master-0 kubenswrapper[10444]: I1205 10:40:38.067053 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:40:38.071918 master-0 kubenswrapper[10444]: I1205 10:40:38.071870 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:40:38.345573 master-0 kubenswrapper[10444]: I1205 10:40:38.345432 10444 generic.go:334] "Generic (PLEG): container finished" podID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerID="8aa1ac52d0f6c8b159f0da42738bae856c9e11f506fdfed545be3d79a9c5ba1d" exitCode=0 Dec 05 10:40:38.346704 master-0 kubenswrapper[10444]: I1205 10:40:38.346357 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" event={"ID":"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0","Type":"ContainerDied","Data":"8aa1ac52d0f6c8b159f0da42738bae856c9e11f506fdfed545be3d79a9c5ba1d"} Dec 05 10:40:38.346906 master-0 kubenswrapper[10444]: I1205 10:40:38.346850 10444 scope.go:117] "RemoveContainer" containerID="8aa1ac52d0f6c8b159f0da42738bae856c9e11f506fdfed545be3d79a9c5ba1d" Dec 05 10:40:38.353658 master-0 kubenswrapper[10444]: I1205 10:40:38.353188 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:40:38.594163 master-0 kubenswrapper[10444]: I1205 10:40:38.594113 10444 scope.go:117] "RemoveContainer" containerID="ded2f19eb6b7e89481ac839fb3ba5dd47ee60cff52debe8dba17401c93d0ac1a" Dec 05 10:40:38.897999 master-0 kubenswrapper[10444]: I1205 10:40:38.897863 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:40:39.354104 master-0 kubenswrapper[10444]: I1205 10:40:39.354040 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5_1e69ce9e-4e6f-4015-9ba6-5a7942570190/openshift-controller-manager-operator/1.log" Dec 05 10:40:39.355071 master-0 kubenswrapper[10444]: I1205 10:40:39.354210 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" event={"ID":"1e69ce9e-4e6f-4015-9ba6-5a7942570190","Type":"ContainerStarted","Data":"d0b2387ade5d83c28cb46bd5046a2e7e9827bca1562cdef34ea49ce95346e6a9"} Dec 05 10:40:39.357523 master-0 kubenswrapper[10444]: I1205 10:40:39.357370 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" event={"ID":"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0","Type":"ContainerStarted","Data":"c3968a3a8ccd5212aba3d71d18db810366b8e5989f94f317e2256250850f0562"} Dec 05 10:40:39.357919 master-0 kubenswrapper[10444]: I1205 10:40:39.357876 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:40:39.362909 master-0 kubenswrapper[10444]: I1205 10:40:39.362832 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:40:39.586374 master-0 kubenswrapper[10444]: I1205 10:40:39.586310 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:40:42.019081 master-0 kubenswrapper[10444]: I1205 10:40:42.019032 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/installer-2-master-0"] Dec 05 10:40:42.019849 master-0 kubenswrapper[10444]: E1205 10:40:42.019828 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6325f9a2-86d5-4fac-9e71-331b2628f965" containerName="installer" Dec 05 10:40:42.019940 master-0 kubenswrapper[10444]: I1205 10:40:42.019926 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="6325f9a2-86d5-4fac-9e71-331b2628f965" containerName="installer" Dec 05 10:40:42.020027 master-0 kubenswrapper[10444]: E1205 10:40:42.020013 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="738d9fe7-2d08-4aa4-842b-9cbeddf120fb" containerName="installer" Dec 05 10:40:42.020095 master-0 kubenswrapper[10444]: I1205 10:40:42.020084 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="738d9fe7-2d08-4aa4-842b-9cbeddf120fb" containerName="installer" Dec 05 10:40:42.020172 master-0 kubenswrapper[10444]: E1205 10:40:42.020161 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e149cbe7-1ec3-4a06-af61-3a8906b8e9ef" containerName="installer" Dec 05 10:40:42.020245 master-0 kubenswrapper[10444]: I1205 10:40:42.020233 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="e149cbe7-1ec3-4a06-af61-3a8906b8e9ef" containerName="installer" Dec 05 10:40:42.020316 master-0 kubenswrapper[10444]: E1205 10:40:42.020304 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d8c854f-31f8-42d3-b28c-82c7d1b14ee9" containerName="installer" Dec 05 10:40:42.020381 master-0 kubenswrapper[10444]: I1205 10:40:42.020371 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d8c854f-31f8-42d3-b28c-82c7d1b14ee9" containerName="installer" Dec 05 10:40:42.020592 master-0 kubenswrapper[10444]: I1205 10:40:42.020576 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d8c854f-31f8-42d3-b28c-82c7d1b14ee9" containerName="installer" Dec 05 10:40:42.020687 master-0 kubenswrapper[10444]: I1205 10:40:42.020675 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="738d9fe7-2d08-4aa4-842b-9cbeddf120fb" containerName="installer" Dec 05 10:40:42.020765 master-0 kubenswrapper[10444]: I1205 10:40:42.020755 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="6325f9a2-86d5-4fac-9e71-331b2628f965" containerName="installer" Dec 05 10:40:42.020846 master-0 kubenswrapper[10444]: I1205 10:40:42.020832 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="e149cbe7-1ec3-4a06-af61-3a8906b8e9ef" containerName="installer" Dec 05 10:40:42.021285 master-0 kubenswrapper[10444]: I1205 10:40:42.021264 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:40:42.023554 master-0 kubenswrapper[10444]: I1205 10:40:42.023504 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-pcdf6" Dec 05 10:40:42.024973 master-0 kubenswrapper[10444]: I1205 10:40:42.024931 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 05 10:40:42.029491 master-0 kubenswrapper[10444]: I1205 10:40:42.029389 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/installer-2-master-0"] Dec 05 10:40:42.196523 master-0 kubenswrapper[10444]: I1205 10:40:42.196394 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:40:42.196788 master-0 kubenswrapper[10444]: I1205 10:40:42.196582 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-var-lock\") pod \"installer-2-master-0\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:40:42.196889 master-0 kubenswrapper[10444]: I1205 10:40:42.196854 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70fbf150-c1f7-4857-8b63-c566282e2526-kube-api-access\") pod \"installer-2-master-0\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:40:42.298307 master-0 kubenswrapper[10444]: I1205 10:40:42.298170 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70fbf150-c1f7-4857-8b63-c566282e2526-kube-api-access\") pod \"installer-2-master-0\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:40:42.298307 master-0 kubenswrapper[10444]: I1205 10:40:42.298296 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:40:42.298926 master-0 kubenswrapper[10444]: I1205 10:40:42.298368 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-var-lock\") pod \"installer-2-master-0\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:40:42.298926 master-0 kubenswrapper[10444]: I1205 10:40:42.298525 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:40:42.298926 master-0 kubenswrapper[10444]: I1205 10:40:42.298561 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-var-lock\") pod \"installer-2-master-0\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:40:42.328447 master-0 kubenswrapper[10444]: I1205 10:40:42.328351 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70fbf150-c1f7-4857-8b63-c566282e2526-kube-api-access\") pod \"installer-2-master-0\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:40:42.388658 master-0 kubenswrapper[10444]: I1205 10:40:42.388592 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:40:42.789666 master-0 kubenswrapper[10444]: I1205 10:40:42.789607 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/installer-2-master-0"] Dec 05 10:40:42.794259 master-0 kubenswrapper[10444]: W1205 10:40:42.794166 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod70fbf150_c1f7_4857_8b63_c566282e2526.slice/crio-452fb60a08a0303dfa4a44b4cff2ac5af4105f5fed0d2d7d9442e582b765e031 WatchSource:0}: Error finding container 452fb60a08a0303dfa4a44b4cff2ac5af4105f5fed0d2d7d9442e582b765e031: Status 404 returned error can't find the container with id 452fb60a08a0303dfa4a44b4cff2ac5af4105f5fed0d2d7d9442e582b765e031 Dec 05 10:40:43.379997 master-0 kubenswrapper[10444]: I1205 10:40:43.379751 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-2-master-0" event={"ID":"70fbf150-c1f7-4857-8b63-c566282e2526","Type":"ContainerStarted","Data":"2215577fb4531349eb7213f364526a3db417316312ccf7c18444e96dbaf859c1"} Dec 05 10:40:43.379997 master-0 kubenswrapper[10444]: I1205 10:40:43.379824 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-2-master-0" event={"ID":"70fbf150-c1f7-4857-8b63-c566282e2526","Type":"ContainerStarted","Data":"452fb60a08a0303dfa4a44b4cff2ac5af4105f5fed0d2d7d9442e582b765e031"} Dec 05 10:40:43.416475 master-0 kubenswrapper[10444]: I1205 10:40:43.415697 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/installer-2-master-0" podStartSLOduration=1.415671562 podStartE2EDuration="1.415671562s" podCreationTimestamp="2025-12-05 10:40:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:40:43.415468697 +0000 UTC m=+184.007381284" watchObservedRunningTime="2025-12-05 10:40:43.415671562 +0000 UTC m=+184.007584149" Dec 05 10:40:48.427045 master-0 kubenswrapper[10444]: I1205 10:40:48.426904 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz"] Dec 05 10:40:48.428314 master-0 kubenswrapper[10444]: I1205 10:40:48.428272 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.429345 master-0 kubenswrapper[10444]: I1205 10:40:48.429272 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f"] Dec 05 10:40:48.430341 master-0 kubenswrapper[10444]: I1205 10:40:48.430301 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:40:48.432024 master-0 kubenswrapper[10444]: I1205 10:40:48.431790 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-88d48b57d-x7jfs"] Dec 05 10:40:48.434372 master-0 kubenswrapper[10444]: I1205 10:40:48.434301 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77"] Dec 05 10:40:48.434654 master-0 kubenswrapper[10444]: I1205 10:40:48.434586 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.435608 master-0 kubenswrapper[10444]: I1205 10:40:48.435550 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-controller-manager-operator"/"cloud-controller-manager-operator-tls" Dec 05 10:40:48.435703 master-0 kubenswrapper[10444]: I1205 10:40:48.435608 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"kube-rbac-proxy" Dec 05 10:40:48.435747 master-0 kubenswrapper[10444]: I1205 10:40:48.435704 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy-cluster-autoscaler-operator" Dec 05 10:40:48.436224 master-0 kubenswrapper[10444]: I1205 10:40:48.436192 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:40:48.436224 master-0 kubenswrapper[10444]: I1205 10:40:48.436215 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 10:40:48.436338 master-0 kubenswrapper[10444]: I1205 10:40:48.436231 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-controller-manager-operator"/"cluster-cloud-controller-manager-dockercfg-45kcs" Dec 05 10:40:48.436387 master-0 kubenswrapper[10444]: I1205 10:40:48.436347 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:48.436455 master-0 kubenswrapper[10444]: I1205 10:40:48.436394 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"cloud-controller-manager-images" Dec 05 10:40:48.436671 master-0 kubenswrapper[10444]: I1205 10:40:48.436640 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-autoscaler-operator-dockercfg-vwmqb" Dec 05 10:40:48.436742 master-0 kubenswrapper[10444]: I1205 10:40:48.436705 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps"] Dec 05 10:40:48.436870 master-0 kubenswrapper[10444]: I1205 10:40:48.436837 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-autoscaler-operator-cert" Dec 05 10:40:48.437293 master-0 kubenswrapper[10444]: I1205 10:40:48.437266 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 10:40:48.437494 master-0 kubenswrapper[10444]: I1205 10:40:48.437466 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:40:48.439475 master-0 kubenswrapper[10444]: I1205 10:40:48.439394 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 10:40:48.439695 master-0 kubenswrapper[10444]: I1205 10:40:48.439670 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 10:40:48.439907 master-0 kubenswrapper[10444]: I1205 10:40:48.439884 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 10:40:48.440494 master-0 kubenswrapper[10444]: I1205 10:40:48.440461 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-insights/insights-operator-55965856b6-2sxv7"] Dec 05 10:40:48.441119 master-0 kubenswrapper[10444]: I1205 10:40:48.441082 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-m8wb4" Dec 05 10:40:48.441233 master-0 kubenswrapper[10444]: I1205 10:40:48.441207 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.441496 master-0 kubenswrapper[10444]: I1205 10:40:48.441413 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-storage-operator"/"cluster-storage-operator-dockercfg-g78vb" Dec 05 10:40:48.441627 master-0 kubenswrapper[10444]: I1205 10:40:48.441598 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-storage-operator"/"cluster-storage-operator-serving-cert" Dec 05 10:40:48.442791 master-0 kubenswrapper[10444]: I1205 10:40:48.442715 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm"] Dec 05 10:40:48.442791 master-0 kubenswrapper[10444]: I1205 10:40:48.442781 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rrnj5" Dec 05 10:40:48.444922 master-0 kubenswrapper[10444]: I1205 10:40:48.444870 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.448109 master-0 kubenswrapper[10444]: I1205 10:40:48.448059 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7"] Dec 05 10:40:48.448725 master-0 kubenswrapper[10444]: I1205 10:40:48.448687 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"trusted-ca-bundle" Dec 05 10:40:48.448806 master-0 kubenswrapper[10444]: I1205 10:40:48.448730 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"openshift-service-ca.crt" Dec 05 10:40:48.449058 master-0 kubenswrapper[10444]: I1205 10:40:48.449023 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:40:48.449214 master-0 kubenswrapper[10444]: I1205 10:40:48.449174 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-baremetal-operator-tls" Dec 05 10:40:48.449678 master-0 kubenswrapper[10444]: I1205 10:40:48.449649 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"kube-root-ca.crt" Dec 05 10:40:48.449754 master-0 kubenswrapper[10444]: I1205 10:40:48.449681 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"service-ca-bundle" Dec 05 10:40:48.449860 master-0 kubenswrapper[10444]: I1205 10:40:48.449831 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-baremetal-operator-dockercfg-h7kbf" Dec 05 10:40:48.449919 master-0 kubenswrapper[10444]: I1205 10:40:48.449899 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-insights"/"openshift-insights-serving-cert" Dec 05 10:40:48.449919 master-0 kubenswrapper[10444]: I1205 10:40:48.449917 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"baremetal-kube-rbac-proxy" Dec 05 10:40:48.450007 master-0 kubenswrapper[10444]: I1205 10:40:48.449919 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-baremetal-webhook-server-cert" Dec 05 10:40:48.450007 master-0 kubenswrapper[10444]: I1205 10:40:48.449964 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-insights"/"operator-dockercfg-lngmk" Dec 05 10:40:48.450169 master-0 kubenswrapper[10444]: I1205 10:40:48.450137 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"cluster-baremetal-operator-images" Dec 05 10:40:48.451479 master-0 kubenswrapper[10444]: I1205 10:40:48.451441 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-kfn9l" Dec 05 10:40:48.451562 master-0 kubenswrapper[10444]: I1205 10:40:48.451526 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 10:40:48.451781 master-0 kubenswrapper[10444]: I1205 10:40:48.451742 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 10:40:48.451846 master-0 kubenswrapper[10444]: I1205 10:40:48.451802 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 10:40:48.458861 master-0 kubenswrapper[10444]: I1205 10:40:48.458797 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p"] Dec 05 10:40:48.459753 master-0 kubenswrapper[10444]: I1205 10:40:48.459723 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:40:48.460485 master-0 kubenswrapper[10444]: I1205 10:40:48.460382 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h"] Dec 05 10:40:48.461288 master-0 kubenswrapper[10444]: I1205 10:40:48.461255 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.462324 master-0 kubenswrapper[10444]: I1205 10:40:48.462284 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-credential-operator"/"cloud-credential-operator-dockercfg-ftghr" Dec 05 10:40:48.463002 master-0 kubenswrapper[10444]: I1205 10:40:48.462948 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-credential-operator"/"cloud-credential-operator-serving-cert" Dec 05 10:40:48.463565 master-0 kubenswrapper[10444]: I1205 10:40:48.463524 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 10:40:48.463902 master-0 kubenswrapper[10444]: I1205 10:40:48.463867 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 10:40:48.463966 master-0 kubenswrapper[10444]: I1205 10:40:48.463912 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-credential-operator"/"kube-root-ca.crt" Dec 05 10:40:48.463966 master-0 kubenswrapper[10444]: I1205 10:40:48.463943 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 10:40:48.464044 master-0 kubenswrapper[10444]: I1205 10:40:48.464009 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-credential-operator"/"openshift-service-ca.crt" Dec 05 10:40:48.464167 master-0 kubenswrapper[10444]: I1205 10:40:48.464140 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-9zg4x" Dec 05 10:40:48.464223 master-0 kubenswrapper[10444]: I1205 10:40:48.464167 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 10:40:48.465346 master-0 kubenswrapper[10444]: I1205 10:40:48.465312 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 10:40:48.470287 master-0 kubenswrapper[10444]: I1205 10:40:48.470214 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c"] Dec 05 10:40:48.470486 master-0 kubenswrapper[10444]: I1205 10:40:48.470329 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-credential-operator"/"cco-trusted-ca" Dec 05 10:40:48.471230 master-0 kubenswrapper[10444]: I1205 10:40:48.471187 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.472865 master-0 kubenswrapper[10444]: I1205 10:40:48.472821 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 10:40:48.477505 master-0 kubenswrapper[10444]: I1205 10:40:48.477367 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f"] Dec 05 10:40:48.498568 master-0 kubenswrapper[10444]: I1205 10:40:48.498218 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p"] Dec 05 10:40:48.507208 master-0 kubenswrapper[10444]: I1205 10:40:48.507157 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7"] Dec 05 10:40:48.507208 master-0 kubenswrapper[10444]: I1205 10:40:48.507207 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-88d48b57d-x7jfs"] Dec 05 10:40:48.511973 master-0 kubenswrapper[10444]: I1205 10:40:48.511911 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps"] Dec 05 10:40:48.511973 master-0 kubenswrapper[10444]: I1205 10:40:48.511975 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77"] Dec 05 10:40:48.515887 master-0 kubenswrapper[10444]: I1205 10:40:48.515827 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h"] Dec 05 10:40:48.517926 master-0 kubenswrapper[10444]: I1205 10:40:48.517787 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c"] Dec 05 10:40:48.534545 master-0 kubenswrapper[10444]: I1205 10:40:48.534463 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-operator-55965856b6-2sxv7"] Dec 05 10:40:48.544861 master-0 kubenswrapper[10444]: I1205 10:40:48.544789 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm"] Dec 05 10:40:48.569079 master-0 kubenswrapper[10444]: I1205 10:40:48.569012 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-config\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.569289 master-0 kubenswrapper[10444]: I1205 10:40:48.569117 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-snapshots\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.569289 master-0 kubenswrapper[10444]: I1205 10:40:48.569152 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cco-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cco-trusted-ca\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:40:48.569289 master-0 kubenswrapper[10444]: I1205 10:40:48.569175 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f69hj\" (UniqueName: \"kubernetes.io/projected/48bd1d86-a6f2-439f-ab04-6a9a442bec42-kube-api-access-f69hj\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.569289 master-0 kubenswrapper[10444]: I1205 10:40:48.569216 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/a73d181f-f4f8-496a-a6f3-8e10804b394f-host-etc-kube\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.569578 master-0 kubenswrapper[10444]: I1205 10:40:48.569521 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cert\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.569578 master-0 kubenswrapper[10444]: I1205 10:40:48.569578 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xjx8\" (UniqueName: \"kubernetes.io/projected/01488ac4-313d-48d5-9e86-7b2011b9e91e-kube-api-access-2xjx8\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.569684 master-0 kubenswrapper[10444]: I1205 10:40:48.569596 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloud-controller-manager-operator-tls\" (UniqueName: \"kubernetes.io/secret/a73d181f-f4f8-496a-a6f3-8e10804b394f-cloud-controller-manager-operator-tls\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.569684 master-0 kubenswrapper[10444]: I1205 10:40:48.569617 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloud-credential-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cloud-credential-operator-serving-cert\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:40:48.569684 master-0 kubenswrapper[10444]: I1205 10:40:48.569634 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-baremetal-operator-tls\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cluster-baremetal-operator-tls\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.569684 master-0 kubenswrapper[10444]: I1205 10:40:48.569654 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fbtj\" (UniqueName: \"kubernetes.io/projected/58b95ae4-7f74-4777-8441-0a0ae28199e9-kube-api-access-7fbtj\") pod \"cluster-storage-operator-dcf7fc84b-9rzps\" (UID: \"58b95ae4-7f74-4777-8441-0a0ae28199e9\") " pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:40:48.569684 master-0 kubenswrapper[10444]: I1205 10:40:48.569669 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/01488ac4-313d-48d5-9e86-7b2011b9e91e-proxy-tls\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.569684 master-0 kubenswrapper[10444]: I1205 10:40:48.569685 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-auth-proxy-config\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.569938 master-0 kubenswrapper[10444]: I1205 10:40:48.569703 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6z8n\" (UniqueName: \"kubernetes.io/projected/4c2975ec-e33d-4960-a708-277d41c79b15-kube-api-access-d6z8n\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:48.569938 master-0 kubenswrapper[10444]: I1205 10:40:48.569754 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ktnb\" (UniqueName: \"kubernetes.io/projected/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-kube-api-access-9ktnb\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:40:48.569938 master-0 kubenswrapper[10444]: I1205 10:40:48.569889 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkzkh\" (UniqueName: \"kubernetes.io/projected/1f132702-304b-46ff-b428-6df1deeffec3-kube-api-access-vkzkh\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:40:48.569938 master-0 kubenswrapper[10444]: I1205 10:40:48.569912 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdc7g\" (UniqueName: \"kubernetes.io/projected/0de26264-46c2-4d29-97da-25a1681d6a8e-kube-api-access-cdc7g\") pod \"cluster-samples-operator-797cfd8b47-glpx7\" (UID: \"0de26264-46c2-4d29-97da-25a1681d6a8e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:40:48.569938 master-0 kubenswrapper[10444]: I1205 10:40:48.569960 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0de26264-46c2-4d29-97da-25a1681d6a8e-samples-operator-tls\") pod \"cluster-samples-operator-797cfd8b47-glpx7\" (UID: \"0de26264-46c2-4d29-97da-25a1681d6a8e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:40:48.569938 master-0 kubenswrapper[10444]: I1205 10:40:48.570016 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f132702-304b-46ff-b428-6df1deeffec3-auth-proxy-config\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:40:48.569938 master-0 kubenswrapper[10444]: I1205 10:40:48.570035 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-webhook-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570081 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-images\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570104 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtsw9\" (UniqueName: \"kubernetes.io/projected/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-kube-api-access-vtsw9\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570120 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8wnp\" (UniqueName: \"kubernetes.io/projected/2c3fab4d-05b8-42ec-8c91-91ff64562649-kube-api-access-l8wnp\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570134 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2c3fab4d-05b8-42ec-8c91-91ff64562649-tmpfs\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570166 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-images\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570181 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-apiservice-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570216 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-profile-collector-cert\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570237 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-auth-proxy-config\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570256 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-storage-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/58b95ae4-7f74-4777-8441-0a0ae28199e9-cluster-storage-operator-serving-cert\") pod \"cluster-storage-operator-dcf7fc84b-9rzps\" (UID: \"58b95ae4-7f74-4777-8441-0a0ae28199e9\") " pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570305 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-serving-cert\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570342 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-trusted-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.570383 master-0 kubenswrapper[10444]: I1205 10:40:48.570387 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/82464e92-4de2-48f0-8772-a489abb16898-machine-api-operator-tls\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.570888 master-0 kubenswrapper[10444]: I1205 10:40:48.570447 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8zp8\" (UniqueName: \"kubernetes.io/projected/a73d181f-f4f8-496a-a6f3-8e10804b394f-kube-api-access-g8zp8\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.570888 master-0 kubenswrapper[10444]: I1205 10:40:48.570470 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f132702-304b-46ff-b428-6df1deeffec3-cert\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:40:48.570888 master-0 kubenswrapper[10444]: I1205 10:40:48.570490 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-srv-cert\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:48.570888 master-0 kubenswrapper[10444]: I1205 10:40:48.570530 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-service-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.570888 master-0 kubenswrapper[10444]: I1205 10:40:48.570563 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dk2b\" (UniqueName: \"kubernetes.io/projected/82464e92-4de2-48f0-8772-a489abb16898-kube-api-access-4dk2b\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.570888 master-0 kubenswrapper[10444]: I1205 10:40:48.570581 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-images\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.570888 master-0 kubenswrapper[10444]: I1205 10:40:48.570597 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-config\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.570888 master-0 kubenswrapper[10444]: I1205 10:40:48.570615 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-images\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.671655 master-0 kubenswrapper[10444]: I1205 10:40:48.671564 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-profile-collector-cert\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:48.671655 master-0 kubenswrapper[10444]: I1205 10:40:48.671644 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-auth-proxy-config\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.671664 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-serving-cert\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.671976 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-storage-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/58b95ae4-7f74-4777-8441-0a0ae28199e9-cluster-storage-operator-serving-cert\") pod \"cluster-storage-operator-dcf7fc84b-9rzps\" (UID: \"58b95ae4-7f74-4777-8441-0a0ae28199e9\") " pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.672021 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-trusted-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.672054 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/82464e92-4de2-48f0-8772-a489abb16898-machine-api-operator-tls\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.672240 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8zp8\" (UniqueName: \"kubernetes.io/projected/a73d181f-f4f8-496a-a6f3-8e10804b394f-kube-api-access-g8zp8\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.672287 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f132702-304b-46ff-b428-6df1deeffec3-cert\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.672324 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-srv-cert\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.672355 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-service-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.672390 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dk2b\" (UniqueName: \"kubernetes.io/projected/82464e92-4de2-48f0-8772-a489abb16898-kube-api-access-4dk2b\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.672413 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-images\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.672838 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-trusted-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.672974 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-images\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673038 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-config\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673086 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-config\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673104 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-images\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673113 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cco-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cco-trusted-ca\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673154 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f69hj\" (UniqueName: \"kubernetes.io/projected/48bd1d86-a6f2-439f-ab04-6a9a442bec42-kube-api-access-f69hj\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673191 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/a73d181f-f4f8-496a-a6f3-8e10804b394f-host-etc-kube\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673217 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-snapshots\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673295 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/a73d181f-f4f8-496a-a6f3-8e10804b394f-host-etc-kube\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673334 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cert\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673364 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xjx8\" (UniqueName: \"kubernetes.io/projected/01488ac4-313d-48d5-9e86-7b2011b9e91e-kube-api-access-2xjx8\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673394 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloud-controller-manager-operator-tls\" (UniqueName: \"kubernetes.io/secret/a73d181f-f4f8-496a-a6f3-8e10804b394f-cloud-controller-manager-operator-tls\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673443 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloud-credential-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cloud-credential-operator-serving-cert\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673470 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-baremetal-operator-tls\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cluster-baremetal-operator-tls\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673538 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/01488ac4-313d-48d5-9e86-7b2011b9e91e-proxy-tls\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673574 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fbtj\" (UniqueName: \"kubernetes.io/projected/58b95ae4-7f74-4777-8441-0a0ae28199e9-kube-api-access-7fbtj\") pod \"cluster-storage-operator-dcf7fc84b-9rzps\" (UID: \"58b95ae4-7f74-4777-8441-0a0ae28199e9\") " pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673622 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-auth-proxy-config\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673649 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6z8n\" (UniqueName: \"kubernetes.io/projected/4c2975ec-e33d-4960-a708-277d41c79b15-kube-api-access-d6z8n\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673677 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ktnb\" (UniqueName: \"kubernetes.io/projected/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-kube-api-access-9ktnb\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673718 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkzkh\" (UniqueName: \"kubernetes.io/projected/1f132702-304b-46ff-b428-6df1deeffec3-kube-api-access-vkzkh\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673746 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdc7g\" (UniqueName: \"kubernetes.io/projected/0de26264-46c2-4d29-97da-25a1681d6a8e-kube-api-access-cdc7g\") pod \"cluster-samples-operator-797cfd8b47-glpx7\" (UID: \"0de26264-46c2-4d29-97da-25a1681d6a8e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673774 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0de26264-46c2-4d29-97da-25a1681d6a8e-samples-operator-tls\") pod \"cluster-samples-operator-797cfd8b47-glpx7\" (UID: \"0de26264-46c2-4d29-97da-25a1681d6a8e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673803 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f132702-304b-46ff-b428-6df1deeffec3-auth-proxy-config\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673808 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-service-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673827 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-webhook-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673884 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-config\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673921 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-images\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.673977 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-images\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.674004 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-snapshots\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.674548 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cco-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cco-trusted-ca\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.674610 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-images\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.674637 master-0 kubenswrapper[10444]: I1205 10:40:48.674673 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtsw9\" (UniqueName: \"kubernetes.io/projected/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-kube-api-access-vtsw9\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.674710 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8wnp\" (UniqueName: \"kubernetes.io/projected/2c3fab4d-05b8-42ec-8c91-91ff64562649-kube-api-access-l8wnp\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.674736 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2c3fab4d-05b8-42ec-8c91-91ff64562649-tmpfs\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.674979 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-images\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.675004 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-apiservice-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.675100 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-auth-proxy-config\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.675694 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2c3fab4d-05b8-42ec-8c91-91ff64562649-tmpfs\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.675959 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-images\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.676118 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-config\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.677703 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f132702-304b-46ff-b428-6df1deeffec3-auth-proxy-config\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.679011 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/82464e92-4de2-48f0-8772-a489abb16898-machine-api-operator-tls\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.680366 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cert\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.688718 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-auth-proxy-config\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.689560 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-webhook-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.690977 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/01488ac4-313d-48d5-9e86-7b2011b9e91e-proxy-tls\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.691059 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloud-controller-manager-operator-tls\" (UniqueName: \"kubernetes.io/secret/a73d181f-f4f8-496a-a6f3-8e10804b394f-cloud-controller-manager-operator-tls\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.691265 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-storage-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/58b95ae4-7f74-4777-8441-0a0ae28199e9-cluster-storage-operator-serving-cert\") pod \"cluster-storage-operator-dcf7fc84b-9rzps\" (UID: \"58b95ae4-7f74-4777-8441-0a0ae28199e9\") " pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.691289 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-apiservice-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.691433 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-serving-cert\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.691526 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0de26264-46c2-4d29-97da-25a1681d6a8e-samples-operator-tls\") pod \"cluster-samples-operator-797cfd8b47-glpx7\" (UID: \"0de26264-46c2-4d29-97da-25a1681d6a8e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.691580 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-srv-cert\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.691843 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-profile-collector-cert\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:48.693021 master-0 kubenswrapper[10444]: I1205 10:40:48.692718 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloud-credential-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cloud-credential-operator-serving-cert\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:40:48.693954 master-0 kubenswrapper[10444]: I1205 10:40:48.693163 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-baremetal-operator-tls\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cluster-baremetal-operator-tls\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.694859 master-0 kubenswrapper[10444]: I1205 10:40:48.694819 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f69hj\" (UniqueName: \"kubernetes.io/projected/48bd1d86-a6f2-439f-ab04-6a9a442bec42-kube-api-access-f69hj\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.695306 master-0 kubenswrapper[10444]: I1205 10:40:48.695268 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6z8n\" (UniqueName: \"kubernetes.io/projected/4c2975ec-e33d-4960-a708-277d41c79b15-kube-api-access-d6z8n\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:48.695633 master-0 kubenswrapper[10444]: I1205 10:40:48.695600 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xjx8\" (UniqueName: \"kubernetes.io/projected/01488ac4-313d-48d5-9e86-7b2011b9e91e-kube-api-access-2xjx8\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.696126 master-0 kubenswrapper[10444]: I1205 10:40:48.696094 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f132702-304b-46ff-b428-6df1deeffec3-cert\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:40:48.696860 master-0 kubenswrapper[10444]: I1205 10:40:48.696189 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fbtj\" (UniqueName: \"kubernetes.io/projected/58b95ae4-7f74-4777-8441-0a0ae28199e9-kube-api-access-7fbtj\") pod \"cluster-storage-operator-dcf7fc84b-9rzps\" (UID: \"58b95ae4-7f74-4777-8441-0a0ae28199e9\") " pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:40:48.696860 master-0 kubenswrapper[10444]: I1205 10:40:48.696340 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkzkh\" (UniqueName: \"kubernetes.io/projected/1f132702-304b-46ff-b428-6df1deeffec3-kube-api-access-vkzkh\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:40:48.696860 master-0 kubenswrapper[10444]: I1205 10:40:48.696346 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8wnp\" (UniqueName: \"kubernetes.io/projected/2c3fab4d-05b8-42ec-8c91-91ff64562649-kube-api-access-l8wnp\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:48.696860 master-0 kubenswrapper[10444]: I1205 10:40:48.696770 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdc7g\" (UniqueName: \"kubernetes.io/projected/0de26264-46c2-4d29-97da-25a1681d6a8e-kube-api-access-cdc7g\") pod \"cluster-samples-operator-797cfd8b47-glpx7\" (UID: \"0de26264-46c2-4d29-97da-25a1681d6a8e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:40:48.696860 master-0 kubenswrapper[10444]: I1205 10:40:48.696805 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ktnb\" (UniqueName: \"kubernetes.io/projected/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-kube-api-access-9ktnb\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:40:48.696860 master-0 kubenswrapper[10444]: I1205 10:40:48.696849 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dk2b\" (UniqueName: \"kubernetes.io/projected/82464e92-4de2-48f0-8772-a489abb16898-kube-api-access-4dk2b\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.697734 master-0 kubenswrapper[10444]: I1205 10:40:48.697694 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtsw9\" (UniqueName: \"kubernetes.io/projected/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-kube-api-access-vtsw9\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.697851 master-0 kubenswrapper[10444]: I1205 10:40:48.697816 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8zp8\" (UniqueName: \"kubernetes.io/projected/a73d181f-f4f8-496a-a6f3-8e10804b394f-kube-api-access-g8zp8\") pod \"cluster-cloud-controller-manager-operator-74f484689c-wn8cz\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.759242 master-0 kubenswrapper[10444]: I1205 10:40:48.759156 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:40:48.773241 master-0 kubenswrapper[10444]: I1205 10:40:48.773175 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:40:48.773241 master-0 kubenswrapper[10444]: W1205 10:40:48.773199 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda73d181f_f4f8_496a_a6f3_8e10804b394f.slice/crio-c7ab07101ebbff7030dc956d8c292ba5a95a74cbbd891a9d901d39cef66dfbf4 WatchSource:0}: Error finding container c7ab07101ebbff7030dc956d8c292ba5a95a74cbbd891a9d901d39cef66dfbf4: Status 404 returned error can't find the container with id c7ab07101ebbff7030dc956d8c292ba5a95a74cbbd891a9d901d39cef66dfbf4 Dec 05 10:40:48.792901 master-0 kubenswrapper[10444]: I1205 10:40:48.792867 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:40:48.807389 master-0 kubenswrapper[10444]: I1205 10:40:48.807339 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:48.851035 master-0 kubenswrapper[10444]: I1205 10:40:48.850972 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:40:48.871152 master-0 kubenswrapper[10444]: I1205 10:40:48.871098 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:40:48.883757 master-0 kubenswrapper[10444]: I1205 10:40:48.883690 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:40:48.889296 master-0 kubenswrapper[10444]: I1205 10:40:48.889250 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:40:48.908281 master-0 kubenswrapper[10444]: I1205 10:40:48.908233 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:40:48.927214 master-0 kubenswrapper[10444]: I1205 10:40:48.927132 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:40:48.938503 master-0 kubenswrapper[10444]: I1205 10:40:48.938451 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:49.184371 master-0 kubenswrapper[10444]: I1205 10:40:49.183337 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f"] Dec 05 10:40:49.253260 master-0 kubenswrapper[10444]: I1205 10:40:49.253165 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-cluster-machine-approver/machine-approver-f797d8546-65t96"] Dec 05 10:40:49.253524 master-0 kubenswrapper[10444]: I1205 10:40:49.253471 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" podUID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" containerName="kube-rbac-proxy" containerID="cri-o://e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608" gracePeriod=30 Dec 05 10:40:49.253771 master-0 kubenswrapper[10444]: I1205 10:40:49.253681 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" podUID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" containerName="machine-approver-controller" containerID="cri-o://764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d" gracePeriod=30 Dec 05 10:40:49.297388 master-0 kubenswrapper[10444]: I1205 10:40:49.292933 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-88d48b57d-x7jfs"] Dec 05 10:40:49.307064 master-0 kubenswrapper[10444]: I1205 10:40:49.306988 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77"] Dec 05 10:40:49.380193 master-0 kubenswrapper[10444]: W1205 10:40:49.380139 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82464e92_4de2_48f0_8772_a489abb16898.slice/crio-897152fdb47d25f0a5016944193be33416b9e61d9cbcab904f0e8c6b63de2a36 WatchSource:0}: Error finding container 897152fdb47d25f0a5016944193be33416b9e61d9cbcab904f0e8c6b63de2a36: Status 404 returned error can't find the container with id 897152fdb47d25f0a5016944193be33416b9e61d9cbcab904f0e8c6b63de2a36 Dec 05 10:40:49.416838 master-0 kubenswrapper[10444]: I1205 10:40:49.416806 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:40:49.419101 master-0 kubenswrapper[10444]: I1205 10:40:49.419044 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" event={"ID":"4c2975ec-e33d-4960-a708-277d41c79b15","Type":"ContainerStarted","Data":"c93da22a302a383f4fdc18e25cf772f0d9d551ca131191ed837accfc6f610c6f"} Dec 05 10:40:49.421797 master-0 kubenswrapper[10444]: I1205 10:40:49.421646 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" event={"ID":"1f132702-304b-46ff-b428-6df1deeffec3","Type":"ContainerStarted","Data":"f12647ed460372c13fb857a5856ac53a5f123d196cc70548c9f47e2b35c490ce"} Dec 05 10:40:49.423400 master-0 kubenswrapper[10444]: I1205 10:40:49.423226 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" event={"ID":"82464e92-4de2-48f0-8772-a489abb16898","Type":"ContainerStarted","Data":"897152fdb47d25f0a5016944193be33416b9e61d9cbcab904f0e8c6b63de2a36"} Dec 05 10:40:49.424356 master-0 kubenswrapper[10444]: I1205 10:40:49.424293 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" event={"ID":"a73d181f-f4f8-496a-a6f3-8e10804b394f","Type":"ContainerStarted","Data":"c7ab07101ebbff7030dc956d8c292ba5a95a74cbbd891a9d901d39cef66dfbf4"} Dec 05 10:40:49.436086 master-0 kubenswrapper[10444]: I1205 10:40:49.436028 10444 generic.go:334] "Generic (PLEG): container finished" podID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" containerID="764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d" exitCode=0 Dec 05 10:40:49.436086 master-0 kubenswrapper[10444]: I1205 10:40:49.436067 10444 generic.go:334] "Generic (PLEG): container finished" podID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" containerID="e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608" exitCode=0 Dec 05 10:40:49.436086 master-0 kubenswrapper[10444]: I1205 10:40:49.436088 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" event={"ID":"e4c3f514-33e8-4deb-932e-56ddf7281c4c","Type":"ContainerDied","Data":"764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d"} Dec 05 10:40:49.436684 master-0 kubenswrapper[10444]: I1205 10:40:49.436104 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" Dec 05 10:40:49.436684 master-0 kubenswrapper[10444]: I1205 10:40:49.436121 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" event={"ID":"e4c3f514-33e8-4deb-932e-56ddf7281c4c","Type":"ContainerDied","Data":"e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608"} Dec 05 10:40:49.436684 master-0 kubenswrapper[10444]: I1205 10:40:49.436132 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-f797d8546-65t96" event={"ID":"e4c3f514-33e8-4deb-932e-56ddf7281c4c","Type":"ContainerDied","Data":"4c18eba53ef1030085a5342e05c911e53f018f33632df043d6b83c7bde56184f"} Dec 05 10:40:49.436684 master-0 kubenswrapper[10444]: I1205 10:40:49.436146 10444 scope.go:117] "RemoveContainer" containerID="764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d" Dec 05 10:40:49.453096 master-0 kubenswrapper[10444]: I1205 10:40:49.453041 10444 scope.go:117] "RemoveContainer" containerID="e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608" Dec 05 10:40:49.475907 master-0 kubenswrapper[10444]: I1205 10:40:49.475795 10444 scope.go:117] "RemoveContainer" containerID="764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d" Dec 05 10:40:49.476370 master-0 kubenswrapper[10444]: E1205 10:40:49.476298 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d\": container with ID starting with 764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d not found: ID does not exist" containerID="764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d" Dec 05 10:40:49.476522 master-0 kubenswrapper[10444]: I1205 10:40:49.476370 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d"} err="failed to get container status \"764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d\": rpc error: code = NotFound desc = could not find container \"764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d\": container with ID starting with 764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d not found: ID does not exist" Dec 05 10:40:49.476522 master-0 kubenswrapper[10444]: I1205 10:40:49.476408 10444 scope.go:117] "RemoveContainer" containerID="e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608" Dec 05 10:40:49.476824 master-0 kubenswrapper[10444]: E1205 10:40:49.476784 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608\": container with ID starting with e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608 not found: ID does not exist" containerID="e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608" Dec 05 10:40:49.476874 master-0 kubenswrapper[10444]: I1205 10:40:49.476818 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608"} err="failed to get container status \"e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608\": rpc error: code = NotFound desc = could not find container \"e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608\": container with ID starting with e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608 not found: ID does not exist" Dec 05 10:40:49.476874 master-0 kubenswrapper[10444]: I1205 10:40:49.476837 10444 scope.go:117] "RemoveContainer" containerID="764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d" Dec 05 10:40:49.477695 master-0 kubenswrapper[10444]: I1205 10:40:49.477489 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d"} err="failed to get container status \"764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d\": rpc error: code = NotFound desc = could not find container \"764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d\": container with ID starting with 764da1adef87c21629d5ea3147a93f78abf3cdffca97a9d1318be4acfa49409d not found: ID does not exist" Dec 05 10:40:49.477695 master-0 kubenswrapper[10444]: I1205 10:40:49.477532 10444 scope.go:117] "RemoveContainer" containerID="e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608" Dec 05 10:40:49.478312 master-0 kubenswrapper[10444]: I1205 10:40:49.478135 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608"} err="failed to get container status \"e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608\": rpc error: code = NotFound desc = could not find container \"e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608\": container with ID starting with e76e5b93c68d92fd9daa3ee4d7f2b34e1a08d418136f3b8db9e7bd4a18875608 not found: ID does not exist" Dec 05 10:40:49.498258 master-0 kubenswrapper[10444]: I1205 10:40:49.498172 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps"] Dec 05 10:40:49.508457 master-0 kubenswrapper[10444]: W1205 10:40:49.508264 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58b95ae4_7f74_4777_8441_0a0ae28199e9.slice/crio-d4750111f1d0e54879f82e850c1814dfc68424e9b8c03f07177ca16103b8a79d WatchSource:0}: Error finding container d4750111f1d0e54879f82e850c1814dfc68424e9b8c03f07177ca16103b8a79d: Status 404 returned error can't find the container with id d4750111f1d0e54879f82e850c1814dfc68424e9b8c03f07177ca16103b8a79d Dec 05 10:40:49.520943 master-0 kubenswrapper[10444]: I1205 10:40:49.520669 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p"] Dec 05 10:40:49.532678 master-0 kubenswrapper[10444]: I1205 10:40:49.532640 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm"] Dec 05 10:40:49.536540 master-0 kubenswrapper[10444]: I1205 10:40:49.536481 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-insights/insights-operator-55965856b6-2sxv7"] Dec 05 10:40:49.552057 master-0 kubenswrapper[10444]: I1205 10:40:49.551807 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h"] Dec 05 10:40:49.601192 master-0 kubenswrapper[10444]: I1205 10:40:49.601149 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nt67z\" (UniqueName: \"kubernetes.io/projected/e4c3f514-33e8-4deb-932e-56ddf7281c4c-kube-api-access-nt67z\") pod \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " Dec 05 10:40:49.601363 master-0 kubenswrapper[10444]: I1205 10:40:49.601230 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-config\") pod \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " Dec 05 10:40:49.601363 master-0 kubenswrapper[10444]: I1205 10:40:49.601291 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-auth-proxy-config\") pod \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " Dec 05 10:40:49.601462 master-0 kubenswrapper[10444]: I1205 10:40:49.601380 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e4c3f514-33e8-4deb-932e-56ddf7281c4c-machine-approver-tls\") pod \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\" (UID: \"e4c3f514-33e8-4deb-932e-56ddf7281c4c\") " Dec 05 10:40:49.603502 master-0 kubenswrapper[10444]: I1205 10:40:49.603458 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-config" (OuterVolumeSpecName: "config") pod "e4c3f514-33e8-4deb-932e-56ddf7281c4c" (UID: "e4c3f514-33e8-4deb-932e-56ddf7281c4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:40:49.604214 master-0 kubenswrapper[10444]: I1205 10:40:49.604182 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "e4c3f514-33e8-4deb-932e-56ddf7281c4c" (UID: "e4c3f514-33e8-4deb-932e-56ddf7281c4c"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:40:49.605055 master-0 kubenswrapper[10444]: I1205 10:40:49.605026 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4c3f514-33e8-4deb-932e-56ddf7281c4c-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "e4c3f514-33e8-4deb-932e-56ddf7281c4c" (UID: "e4c3f514-33e8-4deb-932e-56ddf7281c4c"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:40:49.610460 master-0 kubenswrapper[10444]: I1205 10:40:49.610372 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4c3f514-33e8-4deb-932e-56ddf7281c4c-kube-api-access-nt67z" (OuterVolumeSpecName: "kube-api-access-nt67z") pod "e4c3f514-33e8-4deb-932e-56ddf7281c4c" (UID: "e4c3f514-33e8-4deb-932e-56ddf7281c4c"). InnerVolumeSpecName "kube-api-access-nt67z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:40:49.661108 master-0 kubenswrapper[10444]: I1205 10:40:49.661067 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7"] Dec 05 10:40:49.685518 master-0 kubenswrapper[10444]: I1205 10:40:49.685252 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c"] Dec 05 10:40:49.708690 master-0 kubenswrapper[10444]: I1205 10:40:49.702879 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nt67z\" (UniqueName: \"kubernetes.io/projected/e4c3f514-33e8-4deb-932e-56ddf7281c4c-kube-api-access-nt67z\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:49.708690 master-0 kubenswrapper[10444]: I1205 10:40:49.702928 10444 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:49.708690 master-0 kubenswrapper[10444]: I1205 10:40:49.702941 10444 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e4c3f514-33e8-4deb-932e-56ddf7281c4c-auth-proxy-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:49.708690 master-0 kubenswrapper[10444]: I1205 10:40:49.702973 10444 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/e4c3f514-33e8-4deb-932e-56ddf7281c4c-machine-approver-tls\") on node \"master-0\" DevicePath \"\"" Dec 05 10:40:49.714186 master-0 kubenswrapper[10444]: W1205 10:40:49.714102 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c3fab4d_05b8_42ec_8c91_91ff64562649.slice/crio-6010e9394b879952b424eb35b6227a92d6319988a7bfc5f908a8adc470e9463b WatchSource:0}: Error finding container 6010e9394b879952b424eb35b6227a92d6319988a7bfc5f908a8adc470e9463b: Status 404 returned error can't find the container with id 6010e9394b879952b424eb35b6227a92d6319988a7bfc5f908a8adc470e9463b Dec 05 10:40:49.871967 master-0 kubenswrapper[10444]: I1205 10:40:49.871915 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-cluster-machine-approver/machine-approver-f797d8546-65t96"] Dec 05 10:40:49.876936 master-0 kubenswrapper[10444]: I1205 10:40:49.876837 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-cluster-machine-approver/machine-approver-f797d8546-65t96"] Dec 05 10:40:49.914521 master-0 kubenswrapper[10444]: I1205 10:40:49.914365 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk"] Dec 05 10:40:49.914619 master-0 kubenswrapper[10444]: E1205 10:40:49.914595 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" containerName="machine-approver-controller" Dec 05 10:40:49.914619 master-0 kubenswrapper[10444]: I1205 10:40:49.914608 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" containerName="machine-approver-controller" Dec 05 10:40:49.914619 master-0 kubenswrapper[10444]: E1205 10:40:49.914618 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" containerName="kube-rbac-proxy" Dec 05 10:40:49.914714 master-0 kubenswrapper[10444]: I1205 10:40:49.914624 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" containerName="kube-rbac-proxy" Dec 05 10:40:49.914714 master-0 kubenswrapper[10444]: I1205 10:40:49.914712 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" containerName="machine-approver-controller" Dec 05 10:40:49.914769 master-0 kubenswrapper[10444]: I1205 10:40:49.914730 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" containerName="kube-rbac-proxy" Dec 05 10:40:49.915214 master-0 kubenswrapper[10444]: I1205 10:40:49.915180 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:49.919647 master-0 kubenswrapper[10444]: I1205 10:40:49.918910 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-2qdv2" Dec 05 10:40:49.919647 master-0 kubenswrapper[10444]: I1205 10:40:49.919039 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 10:40:49.919647 master-0 kubenswrapper[10444]: I1205 10:40:49.919383 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 10:40:49.919647 master-0 kubenswrapper[10444]: I1205 10:40:49.919401 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 10:40:49.919647 master-0 kubenswrapper[10444]: I1205 10:40:49.919560 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 10:40:49.919647 master-0 kubenswrapper[10444]: I1205 10:40:49.919625 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 10:40:50.006714 master-0 kubenswrapper[10444]: I1205 10:40:50.006613 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5p8s\" (UniqueName: \"kubernetes.io/projected/62f174ff-9f3c-4051-a34a-e55a59207171-kube-api-access-r5p8s\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.006714 master-0 kubenswrapper[10444]: I1205 10:40:50.006676 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/62f174ff-9f3c-4051-a34a-e55a59207171-machine-approver-tls\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.006714 master-0 kubenswrapper[10444]: I1205 10:40:50.006699 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.006985 master-0 kubenswrapper[10444]: I1205 10:40:50.006937 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-auth-proxy-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.108412 master-0 kubenswrapper[10444]: I1205 10:40:50.108325 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5p8s\" (UniqueName: \"kubernetes.io/projected/62f174ff-9f3c-4051-a34a-e55a59207171-kube-api-access-r5p8s\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.110359 master-0 kubenswrapper[10444]: I1205 10:40:50.108560 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/62f174ff-9f3c-4051-a34a-e55a59207171-machine-approver-tls\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.110359 master-0 kubenswrapper[10444]: I1205 10:40:50.108604 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.110359 master-0 kubenswrapper[10444]: I1205 10:40:50.109170 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-auth-proxy-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.110359 master-0 kubenswrapper[10444]: I1205 10:40:50.109496 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.110359 master-0 kubenswrapper[10444]: I1205 10:40:50.109857 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-auth-proxy-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.115748 master-0 kubenswrapper[10444]: I1205 10:40:50.115686 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/62f174ff-9f3c-4051-a34a-e55a59207171-machine-approver-tls\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.138508 master-0 kubenswrapper[10444]: I1205 10:40:50.138446 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5p8s\" (UniqueName: \"kubernetes.io/projected/62f174ff-9f3c-4051-a34a-e55a59207171-kube-api-access-r5p8s\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.237631 master-0 kubenswrapper[10444]: I1205 10:40:50.237551 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:40:50.274807 master-0 kubenswrapper[10444]: W1205 10:40:50.274358 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62f174ff_9f3c_4051_a34a_e55a59207171.slice/crio-e88afc2116e69a8053c275a20f5cc4b6af9c5241405fc2bf27810f224b5549fb WatchSource:0}: Error finding container e88afc2116e69a8053c275a20f5cc4b6af9c5241405fc2bf27810f224b5549fb: Status 404 returned error can't find the container with id e88afc2116e69a8053c275a20f5cc4b6af9c5241405fc2bf27810f224b5549fb Dec 05 10:40:50.451703 master-0 kubenswrapper[10444]: I1205 10:40:50.451174 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" event={"ID":"01488ac4-313d-48d5-9e86-7b2011b9e91e","Type":"ContainerStarted","Data":"f3885d4c7894400b657866a0c12050353e620b5d55359cad6b836fcd16433ebf"} Dec 05 10:40:50.451703 master-0 kubenswrapper[10444]: I1205 10:40:50.451245 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" event={"ID":"01488ac4-313d-48d5-9e86-7b2011b9e91e","Type":"ContainerStarted","Data":"f9e7a0f7ea008964417b205788660fe8722bce59ff43c7abddbce17f1a9aaf6a"} Dec 05 10:40:50.451703 master-0 kubenswrapper[10444]: I1205 10:40:50.451260 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" event={"ID":"01488ac4-313d-48d5-9e86-7b2011b9e91e","Type":"ContainerStarted","Data":"2cf0d485e10f74fdd9bf0282b8f6fbc341fecfb79764618939e754565cc0b6ff"} Dec 05 10:40:50.466053 master-0 kubenswrapper[10444]: I1205 10:40:50.457164 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" event={"ID":"ee0b4a7f-1b96-4304-bef0-fb575e0e431c","Type":"ContainerStarted","Data":"a1c96c8fbe8416504e81a298ecc76ea203373ea029c15e4f51c2ca83d6b52529"} Dec 05 10:40:50.466053 master-0 kubenswrapper[10444]: I1205 10:40:50.457208 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" event={"ID":"ee0b4a7f-1b96-4304-bef0-fb575e0e431c","Type":"ContainerStarted","Data":"f77a8c44878083fb1f5b837847555bd296fce89d49bff4c5ba5fc4a9953922c2"} Dec 05 10:40:50.466053 master-0 kubenswrapper[10444]: I1205 10:40:50.458333 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" event={"ID":"48bd1d86-a6f2-439f-ab04-6a9a442bec42","Type":"ContainerStarted","Data":"5046caa4441a1760a91554be78c89b5b216c5aae3f10502f03a571a1cc07afd4"} Dec 05 10:40:50.466053 master-0 kubenswrapper[10444]: I1205 10:40:50.460472 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" event={"ID":"1f132702-304b-46ff-b428-6df1deeffec3","Type":"ContainerStarted","Data":"17585b240bf4cdc008578b90ba6cc9fcbae5740e61d0b759e797bb4dea2fc4eb"} Dec 05 10:40:50.466053 master-0 kubenswrapper[10444]: I1205 10:40:50.461673 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" event={"ID":"82464e92-4de2-48f0-8772-a489abb16898","Type":"ContainerStarted","Data":"5078f8bcd22a47f8054039c793714dd265e50535a8498ec8fbafff4e2ddd055d"} Dec 05 10:40:50.471559 master-0 kubenswrapper[10444]: I1205 10:40:50.468606 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" event={"ID":"2c3fab4d-05b8-42ec-8c91-91ff64562649","Type":"ContainerStarted","Data":"4e9a05ab0cfa4d92412fa22fab513f0e72fa6ded6c4964a2d5dc1b99e95f4805"} Dec 05 10:40:50.471559 master-0 kubenswrapper[10444]: I1205 10:40:50.468663 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" event={"ID":"2c3fab4d-05b8-42ec-8c91-91ff64562649","Type":"ContainerStarted","Data":"6010e9394b879952b424eb35b6227a92d6319988a7bfc5f908a8adc470e9463b"} Dec 05 10:40:50.471559 master-0 kubenswrapper[10444]: I1205 10:40:50.471505 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:50.473057 master-0 kubenswrapper[10444]: I1205 10:40:50.472991 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" podStartSLOduration=5.472972928 podStartE2EDuration="5.472972928s" podCreationTimestamp="2025-12-05 10:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:40:50.471994583 +0000 UTC m=+191.063907150" watchObservedRunningTime="2025-12-05 10:40:50.472972928 +0000 UTC m=+191.064885495" Dec 05 10:40:50.475027 master-0 kubenswrapper[10444]: I1205 10:40:50.474892 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-55965856b6-2sxv7" event={"ID":"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd","Type":"ContainerStarted","Data":"c22e02aa77c0fa78cb950ff2c4f24a8244e5ac70b15a797f17b5f262893f0220"} Dec 05 10:40:50.475680 master-0 kubenswrapper[10444]: I1205 10:40:50.475597 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:40:50.477310 master-0 kubenswrapper[10444]: I1205 10:40:50.477271 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" event={"ID":"0de26264-46c2-4d29-97da-25a1681d6a8e","Type":"ContainerStarted","Data":"7c5339e633ef9bd66b822a691a1a94b12e50803e5297f66f5662bfb59a1a2a6c"} Dec 05 10:40:50.479506 master-0 kubenswrapper[10444]: I1205 10:40:50.479462 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" event={"ID":"62f174ff-9f3c-4051-a34a-e55a59207171","Type":"ContainerStarted","Data":"e88afc2116e69a8053c275a20f5cc4b6af9c5241405fc2bf27810f224b5549fb"} Dec 05 10:40:50.481701 master-0 kubenswrapper[10444]: I1205 10:40:50.481662 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" event={"ID":"58b95ae4-7f74-4777-8441-0a0ae28199e9","Type":"ContainerStarted","Data":"d4750111f1d0e54879f82e850c1814dfc68424e9b8c03f07177ca16103b8a79d"} Dec 05 10:40:50.483128 master-0 kubenswrapper[10444]: I1205 10:40:50.483083 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" event={"ID":"4c2975ec-e33d-4960-a708-277d41c79b15","Type":"ContainerStarted","Data":"b40a4d4d28721759b7d6ae727c5d2328ed1b065b72b4a58adbe87d29ca8df256"} Dec 05 10:40:50.483414 master-0 kubenswrapper[10444]: I1205 10:40:50.483359 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:50.491096 master-0 kubenswrapper[10444]: I1205 10:40:50.491035 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:40:50.499127 master-0 kubenswrapper[10444]: I1205 10:40:50.496997 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" podStartSLOduration=5.496978776 podStartE2EDuration="5.496978776s" podCreationTimestamp="2025-12-05 10:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:40:50.494852182 +0000 UTC m=+191.086764769" watchObservedRunningTime="2025-12-05 10:40:50.496978776 +0000 UTC m=+191.088891343" Dec 05 10:40:50.520937 master-0 kubenswrapper[10444]: I1205 10:40:50.520695 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" podStartSLOduration=5.520672596 podStartE2EDuration="5.520672596s" podCreationTimestamp="2025-12-05 10:40:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:40:50.517588217 +0000 UTC m=+191.109500804" watchObservedRunningTime="2025-12-05 10:40:50.520672596 +0000 UTC m=+191.112585163" Dec 05 10:40:50.758727 master-0 kubenswrapper[10444]: I1205 10:40:50.758602 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-djhk8"] Dec 05 10:40:50.759784 master-0 kubenswrapper[10444]: I1205 10:40:50.759741 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:40:50.761884 master-0 kubenswrapper[10444]: I1205 10:40:50.761858 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-fvzgd" Dec 05 10:40:50.770019 master-0 kubenswrapper[10444]: I1205 10:40:50.769953 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-djhk8"] Dec 05 10:40:50.919473 master-0 kubenswrapper[10444]: I1205 10:40:50.919414 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-utilities\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:40:50.919670 master-0 kubenswrapper[10444]: I1205 10:40:50.919513 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pvrn\" (UniqueName: \"kubernetes.io/projected/489b6771-aa16-469f-9883-9a9233d3c379-kube-api-access-9pvrn\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:40:50.919670 master-0 kubenswrapper[10444]: I1205 10:40:50.919575 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-catalog-content\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:40:50.946364 master-0 kubenswrapper[10444]: I1205 10:40:50.946304 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6p8cq"] Dec 05 10:40:50.947569 master-0 kubenswrapper[10444]: I1205 10:40:50.947541 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:40:50.949257 master-0 kubenswrapper[10444]: I1205 10:40:50.949230 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-vp7fw" Dec 05 10:40:50.960489 master-0 kubenswrapper[10444]: I1205 10:40:50.960447 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6p8cq"] Dec 05 10:40:51.022302 master-0 kubenswrapper[10444]: I1205 10:40:51.022136 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-catalog-content\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:40:51.022529 master-0 kubenswrapper[10444]: I1205 10:40:51.022352 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-utilities\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:40:51.022529 master-0 kubenswrapper[10444]: I1205 10:40:51.022451 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pvrn\" (UniqueName: \"kubernetes.io/projected/489b6771-aa16-469f-9883-9a9233d3c379-kube-api-access-9pvrn\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:40:51.025860 master-0 kubenswrapper[10444]: I1205 10:40:51.024124 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-catalog-content\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:40:51.025860 master-0 kubenswrapper[10444]: I1205 10:40:51.024603 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-utilities\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:40:51.046680 master-0 kubenswrapper[10444]: I1205 10:40:51.046633 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pvrn\" (UniqueName: \"kubernetes.io/projected/489b6771-aa16-469f-9883-9a9233d3c379-kube-api-access-9pvrn\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:40:51.116503 master-0 kubenswrapper[10444]: I1205 10:40:51.116454 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:40:51.123322 master-0 kubenswrapper[10444]: I1205 10:40:51.123266 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28qvt\" (UniqueName: \"kubernetes.io/projected/baa812e4-b011-41ee-82ab-8f571f0b7e0a-kube-api-access-28qvt\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:40:51.123490 master-0 kubenswrapper[10444]: I1205 10:40:51.123338 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-utilities\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:40:51.123490 master-0 kubenswrapper[10444]: I1205 10:40:51.123401 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-catalog-content\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:40:51.224991 master-0 kubenswrapper[10444]: I1205 10:40:51.224935 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28qvt\" (UniqueName: \"kubernetes.io/projected/baa812e4-b011-41ee-82ab-8f571f0b7e0a-kube-api-access-28qvt\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:40:51.224991 master-0 kubenswrapper[10444]: I1205 10:40:51.225002 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-utilities\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:40:51.225338 master-0 kubenswrapper[10444]: I1205 10:40:51.225042 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-catalog-content\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:40:51.225510 master-0 kubenswrapper[10444]: I1205 10:40:51.225484 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-catalog-content\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:40:51.232038 master-0 kubenswrapper[10444]: I1205 10:40:51.226082 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-utilities\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:40:51.259513 master-0 kubenswrapper[10444]: I1205 10:40:51.252596 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28qvt\" (UniqueName: \"kubernetes.io/projected/baa812e4-b011-41ee-82ab-8f571f0b7e0a-kube-api-access-28qvt\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:40:51.264372 master-0 kubenswrapper[10444]: I1205 10:40:51.264319 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:40:51.499279 master-0 kubenswrapper[10444]: I1205 10:40:51.499191 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" event={"ID":"62f174ff-9f3c-4051-a34a-e55a59207171","Type":"ContainerStarted","Data":"f445628393287d820027f18b6a0fdccb35344ae04fb12bc19d9b3cf5ae9e9d49"} Dec 05 10:40:51.499279 master-0 kubenswrapper[10444]: I1205 10:40:51.499261 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" event={"ID":"62f174ff-9f3c-4051-a34a-e55a59207171","Type":"ContainerStarted","Data":"3933dffcaf83b19065f79c3940a4d724760bc4d53953aa0eaccc5d23ce8d5d65"} Dec 05 10:40:51.519546 master-0 kubenswrapper[10444]: I1205 10:40:51.519469 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" podStartSLOduration=2.519446264 podStartE2EDuration="2.519446264s" podCreationTimestamp="2025-12-05 10:40:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:40:51.516931039 +0000 UTC m=+192.108843606" watchObservedRunningTime="2025-12-05 10:40:51.519446264 +0000 UTC m=+192.111358831" Dec 05 10:40:51.600843 master-0 kubenswrapper[10444]: I1205 10:40:51.600788 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4c3f514-33e8-4deb-932e-56ddf7281c4c" path="/var/lib/kubelet/pods/e4c3f514-33e8-4deb-932e-56ddf7281c4c/volumes" Dec 05 10:40:52.344389 master-0 kubenswrapper[10444]: I1205 10:40:52.344342 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wk29h"] Dec 05 10:40:52.345268 master-0 kubenswrapper[10444]: I1205 10:40:52.345246 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:40:52.347346 master-0 kubenswrapper[10444]: I1205 10:40:52.347313 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-6j2qd" Dec 05 10:40:52.358454 master-0 kubenswrapper[10444]: I1205 10:40:52.358401 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wk29h"] Dec 05 10:40:52.439317 master-0 kubenswrapper[10444]: I1205 10:40:52.439267 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-utilities\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:40:52.439551 master-0 kubenswrapper[10444]: I1205 10:40:52.439378 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnzgs\" (UniqueName: \"kubernetes.io/projected/0aa6e138-4b1f-4047-8255-a2b14d044588-kube-api-access-wnzgs\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:40:52.439551 master-0 kubenswrapper[10444]: I1205 10:40:52.439462 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-catalog-content\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:40:52.540821 master-0 kubenswrapper[10444]: I1205 10:40:52.540766 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnzgs\" (UniqueName: \"kubernetes.io/projected/0aa6e138-4b1f-4047-8255-a2b14d044588-kube-api-access-wnzgs\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:40:52.540821 master-0 kubenswrapper[10444]: I1205 10:40:52.540843 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-catalog-content\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:40:52.541781 master-0 kubenswrapper[10444]: I1205 10:40:52.540907 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-utilities\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:40:52.541781 master-0 kubenswrapper[10444]: I1205 10:40:52.541395 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-utilities\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:40:52.541781 master-0 kubenswrapper[10444]: I1205 10:40:52.541673 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-catalog-content\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:40:52.558674 master-0 kubenswrapper[10444]: I1205 10:40:52.558634 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnzgs\" (UniqueName: \"kubernetes.io/projected/0aa6e138-4b1f-4047-8255-a2b14d044588-kube-api-access-wnzgs\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:40:52.661512 master-0 kubenswrapper[10444]: I1205 10:40:52.661077 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:40:53.156646 master-0 kubenswrapper[10444]: I1205 10:40:53.156584 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-5n6nw"] Dec 05 10:40:53.157503 master-0 kubenswrapper[10444]: I1205 10:40:53.157476 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.159473 master-0 kubenswrapper[10444]: I1205 10:40:53.159404 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-s9q7m" Dec 05 10:40:53.159601 master-0 kubenswrapper[10444]: I1205 10:40:53.159567 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 10:40:53.165180 master-0 kubenswrapper[10444]: I1205 10:40:53.165118 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-rootfs\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.165358 master-0 kubenswrapper[10444]: I1205 10:40:53.165327 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-mcd-auth-proxy-config\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.165447 master-0 kubenswrapper[10444]: I1205 10:40:53.165377 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-proxy-tls\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.165447 master-0 kubenswrapper[10444]: I1205 10:40:53.165405 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvw27\" (UniqueName: \"kubernetes.io/projected/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-kube-api-access-rvw27\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.266497 master-0 kubenswrapper[10444]: I1205 10:40:53.266441 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-mcd-auth-proxy-config\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.266497 master-0 kubenswrapper[10444]: I1205 10:40:53.266506 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-proxy-tls\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.266949 master-0 kubenswrapper[10444]: I1205 10:40:53.266918 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvw27\" (UniqueName: \"kubernetes.io/projected/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-kube-api-access-rvw27\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.267031 master-0 kubenswrapper[10444]: I1205 10:40:53.267014 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-rootfs\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.267119 master-0 kubenswrapper[10444]: I1205 10:40:53.267105 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-rootfs\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.267824 master-0 kubenswrapper[10444]: I1205 10:40:53.267805 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-mcd-auth-proxy-config\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.271150 master-0 kubenswrapper[10444]: I1205 10:40:53.271128 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-proxy-tls\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.285411 master-0 kubenswrapper[10444]: I1205 10:40:53.283749 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvw27\" (UniqueName: \"kubernetes.io/projected/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-kube-api-access-rvw27\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.481874 master-0 kubenswrapper[10444]: I1205 10:40:53.481725 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:40:53.551964 master-0 kubenswrapper[10444]: I1205 10:40:53.551904 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pqhfn"] Dec 05 10:40:53.554084 master-0 kubenswrapper[10444]: I1205 10:40:53.554061 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:40:53.556323 master-0 kubenswrapper[10444]: I1205 10:40:53.556291 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-qq76p" Dec 05 10:40:53.563280 master-0 kubenswrapper[10444]: I1205 10:40:53.563195 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pqhfn"] Dec 05 10:40:53.672307 master-0 kubenswrapper[10444]: I1205 10:40:53.671798 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4jgc\" (UniqueName: \"kubernetes.io/projected/0a4db531-8af7-4085-a6b2-0de51b527ce6-kube-api-access-g4jgc\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:40:53.672307 master-0 kubenswrapper[10444]: I1205 10:40:53.672047 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-utilities\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:40:53.672307 master-0 kubenswrapper[10444]: I1205 10:40:53.672070 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-catalog-content\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:40:53.773813 master-0 kubenswrapper[10444]: I1205 10:40:53.773596 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4jgc\" (UniqueName: \"kubernetes.io/projected/0a4db531-8af7-4085-a6b2-0de51b527ce6-kube-api-access-g4jgc\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:40:53.774165 master-0 kubenswrapper[10444]: I1205 10:40:53.774104 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-utilities\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:40:53.774255 master-0 kubenswrapper[10444]: I1205 10:40:53.774199 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-utilities\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:40:53.774448 master-0 kubenswrapper[10444]: I1205 10:40:53.774360 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-catalog-content\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:40:53.774791 master-0 kubenswrapper[10444]: I1205 10:40:53.774727 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-catalog-content\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:40:53.791849 master-0 kubenswrapper[10444]: I1205 10:40:53.791796 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4jgc\" (UniqueName: \"kubernetes.io/projected/0a4db531-8af7-4085-a6b2-0de51b527ce6-kube-api-access-g4jgc\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:40:53.878451 master-0 kubenswrapper[10444]: I1205 10:40:53.878377 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:40:55.357440 master-0 kubenswrapper[10444]: I1205 10:40:55.357364 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz"] Dec 05 10:41:03.591160 master-0 kubenswrapper[10444]: I1205 10:41:03.590818 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" event={"ID":"0de26264-46c2-4d29-97da-25a1681d6a8e","Type":"ContainerStarted","Data":"17b13bc43c4778d13fa687a31758ca2dbca56588642f33777f983dd690e3b116"} Dec 05 10:41:03.608758 master-0 kubenswrapper[10444]: I1205 10:41:03.608702 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" event={"ID":"b7e7f216-f9af-41df-a0b0-df2a76b9b72a","Type":"ContainerStarted","Data":"07cb2a5b59f4a1cf9ae60c0a62bf4e44ad34c98cbd92efe399d17d750ec292c6"} Dec 05 10:41:03.865335 master-0 kubenswrapper[10444]: I1205 10:41:03.865264 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wk29h"] Dec 05 10:41:03.881793 master-0 kubenswrapper[10444]: I1205 10:41:03.878658 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-djhk8"] Dec 05 10:41:03.881793 master-0 kubenswrapper[10444]: W1205 10:41:03.880048 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0aa6e138_4b1f_4047_8255_a2b14d044588.slice/crio-52b24a28414c04f2574070f038561f387ec44118f13f6cac5eab1761c0750f3a WatchSource:0}: Error finding container 52b24a28414c04f2574070f038561f387ec44118f13f6cac5eab1761c0750f3a: Status 404 returned error can't find the container with id 52b24a28414c04f2574070f038561f387ec44118f13f6cac5eab1761c0750f3a Dec 05 10:41:03.882007 master-0 kubenswrapper[10444]: I1205 10:41:03.881947 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6p8cq"] Dec 05 10:41:03.883683 master-0 kubenswrapper[10444]: W1205 10:41:03.883595 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbaa812e4_b011_41ee_82ab_8f571f0b7e0a.slice/crio-35b4f066f81dc683769e8a58a956f46b65324a32b3270c15056ef67e541a8fdd WatchSource:0}: Error finding container 35b4f066f81dc683769e8a58a956f46b65324a32b3270c15056ef67e541a8fdd: Status 404 returned error can't find the container with id 35b4f066f81dc683769e8a58a956f46b65324a32b3270c15056ef67e541a8fdd Dec 05 10:41:04.317362 master-0 kubenswrapper[10444]: I1205 10:41:04.317279 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pqhfn"] Dec 05 10:41:04.601767 master-0 kubenswrapper[10444]: I1205 10:41:04.601720 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" event={"ID":"1f132702-304b-46ff-b428-6df1deeffec3","Type":"ContainerStarted","Data":"a2f6a915977700175ebb2e8af148b97bcb382c07417e2ec1c7af394d18c91e58"} Dec 05 10:41:04.615850 master-0 kubenswrapper[10444]: I1205 10:41:04.606581 10444 generic.go:334] "Generic (PLEG): container finished" podID="baa812e4-b011-41ee-82ab-8f571f0b7e0a" containerID="b0504c32d8948d61de04b9482cfad535351eb3152188b508b07ba9be49d98d54" exitCode=0 Dec 05 10:41:04.615850 master-0 kubenswrapper[10444]: I1205 10:41:04.606654 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6p8cq" event={"ID":"baa812e4-b011-41ee-82ab-8f571f0b7e0a","Type":"ContainerDied","Data":"b0504c32d8948d61de04b9482cfad535351eb3152188b508b07ba9be49d98d54"} Dec 05 10:41:04.615850 master-0 kubenswrapper[10444]: I1205 10:41:04.606683 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6p8cq" event={"ID":"baa812e4-b011-41ee-82ab-8f571f0b7e0a","Type":"ContainerStarted","Data":"35b4f066f81dc683769e8a58a956f46b65324a32b3270c15056ef67e541a8fdd"} Dec 05 10:41:04.615850 master-0 kubenswrapper[10444]: I1205 10:41:04.608878 10444 generic.go:334] "Generic (PLEG): container finished" podID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerID="698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90" exitCode=0 Dec 05 10:41:04.615850 master-0 kubenswrapper[10444]: I1205 10:41:04.608988 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk29h" event={"ID":"0aa6e138-4b1f-4047-8255-a2b14d044588","Type":"ContainerDied","Data":"698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90"} Dec 05 10:41:04.615850 master-0 kubenswrapper[10444]: I1205 10:41:04.609064 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk29h" event={"ID":"0aa6e138-4b1f-4047-8255-a2b14d044588","Type":"ContainerStarted","Data":"52b24a28414c04f2574070f038561f387ec44118f13f6cac5eab1761c0750f3a"} Dec 05 10:41:04.615850 master-0 kubenswrapper[10444]: I1205 10:41:04.610237 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqhfn" event={"ID":"0a4db531-8af7-4085-a6b2-0de51b527ce6","Type":"ContainerStarted","Data":"8904f4db7f08be6823cd5712612b6b29bb4d4506ff20a7f693a2b99016d20918"} Dec 05 10:41:04.616784 master-0 kubenswrapper[10444]: I1205 10:41:04.616688 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-insights/insights-operator-55965856b6-2sxv7" event={"ID":"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd","Type":"ContainerStarted","Data":"988521f321f76ee1338b66d529084336cd28c9aff898aebeb3c9898e6eee09c2"} Dec 05 10:41:04.622478 master-0 kubenswrapper[10444]: I1205 10:41:04.622402 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" event={"ID":"0de26264-46c2-4d29-97da-25a1681d6a8e","Type":"ContainerStarted","Data":"e85483699ad53d56514029bdfb8da39288932c07477e524338d3b74635fb69cb"} Dec 05 10:41:04.630504 master-0 kubenswrapper[10444]: I1205 10:41:04.629575 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" event={"ID":"48bd1d86-a6f2-439f-ab04-6a9a442bec42","Type":"ContainerStarted","Data":"202d02bb3e4ad58d66f99961bf845501eb1c6c405b6931ea76d28b61eec60c5e"} Dec 05 10:41:04.630504 master-0 kubenswrapper[10444]: I1205 10:41:04.629642 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" event={"ID":"48bd1d86-a6f2-439f-ab04-6a9a442bec42","Type":"ContainerStarted","Data":"3a9cca203c845359b6a28fb821277d5c1db034d15078e96e3ff87b488109d4de"} Dec 05 10:41:04.631803 master-0 kubenswrapper[10444]: I1205 10:41:04.631710 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" event={"ID":"58b95ae4-7f74-4777-8441-0a0ae28199e9","Type":"ContainerStarted","Data":"b4f5e2b15882f1260cb91c5c5fa464782c9f05a037d4df711995ab32dc4b7d1b"} Dec 05 10:41:04.637501 master-0 kubenswrapper[10444]: I1205 10:41:04.637329 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" podStartSLOduration=5.822357904 podStartE2EDuration="19.637301301s" podCreationTimestamp="2025-12-05 10:40:45 +0000 UTC" firstStartedPulling="2025-12-05 10:40:49.513830281 +0000 UTC m=+190.105742848" lastFinishedPulling="2025-12-05 10:41:03.328773678 +0000 UTC m=+203.920686245" observedRunningTime="2025-12-05 10:41:04.628954897 +0000 UTC m=+205.220867464" watchObservedRunningTime="2025-12-05 10:41:04.637301301 +0000 UTC m=+205.229213888" Dec 05 10:41:04.714521 master-0 kubenswrapper[10444]: I1205 10:41:04.701835 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" event={"ID":"b7e7f216-f9af-41df-a0b0-df2a76b9b72a","Type":"ContainerStarted","Data":"a0c6598bae970a093387ebf1da6d639ef058d5bc4cc4461c9eae0e3d29fec097"} Dec 05 10:41:04.714521 master-0 kubenswrapper[10444]: I1205 10:41:04.701920 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" event={"ID":"b7e7f216-f9af-41df-a0b0-df2a76b9b72a","Type":"ContainerStarted","Data":"fd4950aba5a49e315ad2c96cee0227a74e29a98515c23eef501653ae60bc53c2"} Dec 05 10:41:04.714521 master-0 kubenswrapper[10444]: I1205 10:41:04.707551 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" event={"ID":"82464e92-4de2-48f0-8772-a489abb16898","Type":"ContainerStarted","Data":"24a5089a50a474ab31ec97627465524449e5c82bd6525e30a758626b2ffde4b6"} Dec 05 10:41:04.714521 master-0 kubenswrapper[10444]: I1205 10:41:04.710712 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" event={"ID":"a73d181f-f4f8-496a-a6f3-8e10804b394f","Type":"ContainerStarted","Data":"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83"} Dec 05 10:41:04.714521 master-0 kubenswrapper[10444]: I1205 10:41:04.710759 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" event={"ID":"a73d181f-f4f8-496a-a6f3-8e10804b394f","Type":"ContainerStarted","Data":"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05"} Dec 05 10:41:04.714521 master-0 kubenswrapper[10444]: I1205 10:41:04.713085 10444 generic.go:334] "Generic (PLEG): container finished" podID="489b6771-aa16-469f-9883-9a9233d3c379" containerID="8338111de9d79599f0bf5e87a8b5d537443ba25abd595ae9a7a16a59ac9bb008" exitCode=0 Dec 05 10:41:04.714521 master-0 kubenswrapper[10444]: I1205 10:41:04.713175 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djhk8" event={"ID":"489b6771-aa16-469f-9883-9a9233d3c379","Type":"ContainerDied","Data":"8338111de9d79599f0bf5e87a8b5d537443ba25abd595ae9a7a16a59ac9bb008"} Dec 05 10:41:04.714521 master-0 kubenswrapper[10444]: I1205 10:41:04.713200 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djhk8" event={"ID":"489b6771-aa16-469f-9883-9a9233d3c379","Type":"ContainerStarted","Data":"4c76b1c660e01ccc1ed121db81862ca5648ff5a628cbb72bf8b46891908897a5"} Dec 05 10:41:04.720292 master-0 kubenswrapper[10444]: I1205 10:41:04.718140 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" event={"ID":"ee0b4a7f-1b96-4304-bef0-fb575e0e431c","Type":"ContainerStarted","Data":"3ff68c3e87818d7e554f21939713ccda5dbb7fbfa8357657564f64a7cbb6eb58"} Dec 05 10:41:04.764104 master-0 kubenswrapper[10444]: I1205 10:41:04.764012 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" podStartSLOduration=7.294614774 podStartE2EDuration="20.763992223s" podCreationTimestamp="2025-12-05 10:40:44 +0000 UTC" firstStartedPulling="2025-12-05 10:40:49.882466814 +0000 UTC m=+190.474379391" lastFinishedPulling="2025-12-05 10:41:03.351844273 +0000 UTC m=+203.943756840" observedRunningTime="2025-12-05 10:41:04.762577287 +0000 UTC m=+205.354489854" watchObservedRunningTime="2025-12-05 10:41:04.763992223 +0000 UTC m=+205.355904790" Dec 05 10:41:04.789719 master-0 kubenswrapper[10444]: I1205 10:41:04.787556 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-insights/insights-operator-55965856b6-2sxv7" podStartSLOduration=7.008571768 podStartE2EDuration="20.787532929s" podCreationTimestamp="2025-12-05 10:40:44 +0000 UTC" firstStartedPulling="2025-12-05 10:40:49.554754335 +0000 UTC m=+190.146666892" lastFinishedPulling="2025-12-05 10:41:03.333715486 +0000 UTC m=+203.925628053" observedRunningTime="2025-12-05 10:41:04.783926666 +0000 UTC m=+205.375839233" watchObservedRunningTime="2025-12-05 10:41:04.787532929 +0000 UTC m=+205.379445516" Dec 05 10:41:04.844253 master-0 kubenswrapper[10444]: I1205 10:41:04.843966 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" podStartSLOduration=6.046823783 podStartE2EDuration="19.843946752s" podCreationTimestamp="2025-12-05 10:40:45 +0000 UTC" firstStartedPulling="2025-12-05 10:40:49.554740924 +0000 UTC m=+190.146653491" lastFinishedPulling="2025-12-05 10:41:03.351863893 +0000 UTC m=+203.943776460" observedRunningTime="2025-12-05 10:41:04.816900016 +0000 UTC m=+205.408812583" watchObservedRunningTime="2025-12-05 10:41:04.843946752 +0000 UTC m=+205.435859329" Dec 05 10:41:04.845689 master-0 kubenswrapper[10444]: I1205 10:41:04.845643 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" podStartSLOduration=5.982054776 podStartE2EDuration="19.845635095s" podCreationTimestamp="2025-12-05 10:40:45 +0000 UTC" firstStartedPulling="2025-12-05 10:40:49.571703352 +0000 UTC m=+190.163615919" lastFinishedPulling="2025-12-05 10:41:03.435283671 +0000 UTC m=+204.027196238" observedRunningTime="2025-12-05 10:41:04.843790488 +0000 UTC m=+205.435703055" watchObservedRunningTime="2025-12-05 10:41:04.845635095 +0000 UTC m=+205.437547662" Dec 05 10:41:04.874750 master-0 kubenswrapper[10444]: I1205 10:41:04.874681 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" podStartSLOduration=6.997760981 podStartE2EDuration="20.874666543s" podCreationTimestamp="2025-12-05 10:40:44 +0000 UTC" firstStartedPulling="2025-12-05 10:40:49.513594595 +0000 UTC m=+190.105507162" lastFinishedPulling="2025-12-05 10:41:03.390500157 +0000 UTC m=+203.982412724" observedRunningTime="2025-12-05 10:41:04.873590065 +0000 UTC m=+205.465502652" watchObservedRunningTime="2025-12-05 10:41:04.874666543 +0000 UTC m=+205.466579110" Dec 05 10:41:04.923930 master-0 kubenswrapper[10444]: I1205 10:41:04.923216 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" podStartSLOduration=11.923189153 podStartE2EDuration="11.923189153s" podCreationTimestamp="2025-12-05 10:40:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:41:04.918632145 +0000 UTC m=+205.510544722" watchObservedRunningTime="2025-12-05 10:41:04.923189153 +0000 UTC m=+205.515101720" Dec 05 10:41:04.950517 master-0 kubenswrapper[10444]: I1205 10:41:04.949253 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" podStartSLOduration=7.248117166 podStartE2EDuration="20.949228382s" podCreationTimestamp="2025-12-05 10:40:44 +0000 UTC" firstStartedPulling="2025-12-05 10:40:49.69281392 +0000 UTC m=+190.284726487" lastFinishedPulling="2025-12-05 10:41:03.393925136 +0000 UTC m=+203.985837703" observedRunningTime="2025-12-05 10:41:04.940810116 +0000 UTC m=+205.532722683" watchObservedRunningTime="2025-12-05 10:41:04.949228382 +0000 UTC m=+205.541140979" Dec 05 10:41:05.726545 master-0 kubenswrapper[10444]: I1205 10:41:05.726480 10444 generic.go:334] "Generic (PLEG): container finished" podID="0a4db531-8af7-4085-a6b2-0de51b527ce6" containerID="b5d5f4343a07894afd8f108c3807a5c794b99b06442a88e2827b3d948daa8970" exitCode=0 Dec 05 10:41:05.727207 master-0 kubenswrapper[10444]: I1205 10:41:05.726574 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqhfn" event={"ID":"0a4db531-8af7-4085-a6b2-0de51b527ce6","Type":"ContainerDied","Data":"b5d5f4343a07894afd8f108c3807a5c794b99b06442a88e2827b3d948daa8970"} Dec 05 10:41:05.729026 master-0 kubenswrapper[10444]: I1205 10:41:05.729001 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" event={"ID":"a73d181f-f4f8-496a-a6f3-8e10804b394f","Type":"ContainerStarted","Data":"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c"} Dec 05 10:41:05.730338 master-0 kubenswrapper[10444]: I1205 10:41:05.730187 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="cluster-cloud-controller-manager" containerID="cri-o://29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05" gracePeriod=30 Dec 05 10:41:05.730338 master-0 kubenswrapper[10444]: I1205 10:41:05.730217 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="kube-rbac-proxy" containerID="cri-o://9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c" gracePeriod=30 Dec 05 10:41:05.730338 master-0 kubenswrapper[10444]: I1205 10:41:05.730252 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="config-sync-controllers" containerID="cri-o://b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83" gracePeriod=30 Dec 05 10:41:05.772629 master-0 kubenswrapper[10444]: I1205 10:41:05.772529 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" podStartSLOduration=7.221890682 podStartE2EDuration="21.772510971s" podCreationTimestamp="2025-12-05 10:40:44 +0000 UTC" firstStartedPulling="2025-12-05 10:40:48.782953802 +0000 UTC m=+189.374866369" lastFinishedPulling="2025-12-05 10:41:03.333574091 +0000 UTC m=+203.925486658" observedRunningTime="2025-12-05 10:41:05.770606282 +0000 UTC m=+206.362518839" watchObservedRunningTime="2025-12-05 10:41:05.772510971 +0000 UTC m=+206.364423528" Dec 05 10:41:05.900835 master-0 kubenswrapper[10444]: I1205 10:41:05.900761 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:41:06.050090 master-0 kubenswrapper[10444]: I1205 10:41:06.050027 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-images\") pod \"a73d181f-f4f8-496a-a6f3-8e10804b394f\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " Dec 05 10:41:06.050090 master-0 kubenswrapper[10444]: I1205 10:41:06.050090 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8zp8\" (UniqueName: \"kubernetes.io/projected/a73d181f-f4f8-496a-a6f3-8e10804b394f-kube-api-access-g8zp8\") pod \"a73d181f-f4f8-496a-a6f3-8e10804b394f\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " Dec 05 10:41:06.050354 master-0 kubenswrapper[10444]: I1205 10:41:06.050152 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-auth-proxy-config\") pod \"a73d181f-f4f8-496a-a6f3-8e10804b394f\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " Dec 05 10:41:06.050354 master-0 kubenswrapper[10444]: I1205 10:41:06.050198 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/a73d181f-f4f8-496a-a6f3-8e10804b394f-host-etc-kube\") pod \"a73d181f-f4f8-496a-a6f3-8e10804b394f\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " Dec 05 10:41:06.050354 master-0 kubenswrapper[10444]: I1205 10:41:06.050237 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cloud-controller-manager-operator-tls\" (UniqueName: \"kubernetes.io/secret/a73d181f-f4f8-496a-a6f3-8e10804b394f-cloud-controller-manager-operator-tls\") pod \"a73d181f-f4f8-496a-a6f3-8e10804b394f\" (UID: \"a73d181f-f4f8-496a-a6f3-8e10804b394f\") " Dec 05 10:41:06.050513 master-0 kubenswrapper[10444]: I1205 10:41:06.050347 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a73d181f-f4f8-496a-a6f3-8e10804b394f-host-etc-kube" (OuterVolumeSpecName: "host-etc-kube") pod "a73d181f-f4f8-496a-a6f3-8e10804b394f" (UID: "a73d181f-f4f8-496a-a6f3-8e10804b394f"). InnerVolumeSpecName "host-etc-kube". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:41:06.050745 master-0 kubenswrapper[10444]: I1205 10:41:06.050717 10444 reconciler_common.go:293] "Volume detached for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/a73d181f-f4f8-496a-a6f3-8e10804b394f-host-etc-kube\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:06.051171 master-0 kubenswrapper[10444]: I1205 10:41:06.051119 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "a73d181f-f4f8-496a-a6f3-8e10804b394f" (UID: "a73d181f-f4f8-496a-a6f3-8e10804b394f"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:41:06.051237 master-0 kubenswrapper[10444]: I1205 10:41:06.051134 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-images" (OuterVolumeSpecName: "images") pod "a73d181f-f4f8-496a-a6f3-8e10804b394f" (UID: "a73d181f-f4f8-496a-a6f3-8e10804b394f"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:41:06.054341 master-0 kubenswrapper[10444]: I1205 10:41:06.054307 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a73d181f-f4f8-496a-a6f3-8e10804b394f-cloud-controller-manager-operator-tls" (OuterVolumeSpecName: "cloud-controller-manager-operator-tls") pod "a73d181f-f4f8-496a-a6f3-8e10804b394f" (UID: "a73d181f-f4f8-496a-a6f3-8e10804b394f"). InnerVolumeSpecName "cloud-controller-manager-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:41:06.067692 master-0 kubenswrapper[10444]: I1205 10:41:06.067637 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a73d181f-f4f8-496a-a6f3-8e10804b394f-kube-api-access-g8zp8" (OuterVolumeSpecName: "kube-api-access-g8zp8") pod "a73d181f-f4f8-496a-a6f3-8e10804b394f" (UID: "a73d181f-f4f8-496a-a6f3-8e10804b394f"). InnerVolumeSpecName "kube-api-access-g8zp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:41:06.151840 master-0 kubenswrapper[10444]: I1205 10:41:06.151779 10444 reconciler_common.go:293] "Volume detached for volume \"cloud-controller-manager-operator-tls\" (UniqueName: \"kubernetes.io/secret/a73d181f-f4f8-496a-a6f3-8e10804b394f-cloud-controller-manager-operator-tls\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:06.151840 master-0 kubenswrapper[10444]: I1205 10:41:06.151835 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8zp8\" (UniqueName: \"kubernetes.io/projected/a73d181f-f4f8-496a-a6f3-8e10804b394f-kube-api-access-g8zp8\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:06.151840 master-0 kubenswrapper[10444]: I1205 10:41:06.151853 10444 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-images\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:06.152142 master-0 kubenswrapper[10444]: I1205 10:41:06.151866 10444 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a73d181f-f4f8-496a-a6f3-8e10804b394f-auth-proxy-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:06.740618 master-0 kubenswrapper[10444]: I1205 10:41:06.740531 10444 generic.go:334] "Generic (PLEG): container finished" podID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerID="9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c" exitCode=0 Dec 05 10:41:06.740618 master-0 kubenswrapper[10444]: I1205 10:41:06.740580 10444 generic.go:334] "Generic (PLEG): container finished" podID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerID="b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83" exitCode=0 Dec 05 10:41:06.740618 master-0 kubenswrapper[10444]: I1205 10:41:06.740591 10444 generic.go:334] "Generic (PLEG): container finished" podID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerID="29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05" exitCode=0 Dec 05 10:41:06.740618 master-0 kubenswrapper[10444]: I1205 10:41:06.740614 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" event={"ID":"a73d181f-f4f8-496a-a6f3-8e10804b394f","Type":"ContainerDied","Data":"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c"} Dec 05 10:41:06.741829 master-0 kubenswrapper[10444]: I1205 10:41:06.740653 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" event={"ID":"a73d181f-f4f8-496a-a6f3-8e10804b394f","Type":"ContainerDied","Data":"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83"} Dec 05 10:41:06.741829 master-0 kubenswrapper[10444]: I1205 10:41:06.740668 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" event={"ID":"a73d181f-f4f8-496a-a6f3-8e10804b394f","Type":"ContainerDied","Data":"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05"} Dec 05 10:41:06.741829 master-0 kubenswrapper[10444]: I1205 10:41:06.740684 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" event={"ID":"a73d181f-f4f8-496a-a6f3-8e10804b394f","Type":"ContainerDied","Data":"c7ab07101ebbff7030dc956d8c292ba5a95a74cbbd891a9d901d39cef66dfbf4"} Dec 05 10:41:06.741829 master-0 kubenswrapper[10444]: I1205 10:41:06.740704 10444 scope.go:117] "RemoveContainer" containerID="9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c" Dec 05 10:41:06.741829 master-0 kubenswrapper[10444]: I1205 10:41:06.740800 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz" Dec 05 10:41:06.797837 master-0 kubenswrapper[10444]: I1205 10:41:06.797465 10444 scope.go:117] "RemoveContainer" containerID="b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83" Dec 05 10:41:06.805514 master-0 kubenswrapper[10444]: I1205 10:41:06.805472 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz"] Dec 05 10:41:06.820467 master-0 kubenswrapper[10444]: I1205 10:41:06.820360 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-74f484689c-wn8cz"] Dec 05 10:41:06.822723 master-0 kubenswrapper[10444]: I1205 10:41:06.822615 10444 scope.go:117] "RemoveContainer" containerID="29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05" Dec 05 10:41:06.841055 master-0 kubenswrapper[10444]: I1205 10:41:06.838764 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz"] Dec 05 10:41:06.841055 master-0 kubenswrapper[10444]: E1205 10:41:06.839042 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="cluster-cloud-controller-manager" Dec 05 10:41:06.841055 master-0 kubenswrapper[10444]: I1205 10:41:06.839059 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="cluster-cloud-controller-manager" Dec 05 10:41:06.841055 master-0 kubenswrapper[10444]: E1205 10:41:06.839075 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="config-sync-controllers" Dec 05 10:41:06.841055 master-0 kubenswrapper[10444]: I1205 10:41:06.839083 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="config-sync-controllers" Dec 05 10:41:06.841055 master-0 kubenswrapper[10444]: E1205 10:41:06.839099 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="kube-rbac-proxy" Dec 05 10:41:06.841055 master-0 kubenswrapper[10444]: I1205 10:41:06.839107 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="kube-rbac-proxy" Dec 05 10:41:06.841055 master-0 kubenswrapper[10444]: I1205 10:41:06.839223 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="config-sync-controllers" Dec 05 10:41:06.841055 master-0 kubenswrapper[10444]: I1205 10:41:06.839241 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="kube-rbac-proxy" Dec 05 10:41:06.841055 master-0 kubenswrapper[10444]: I1205 10:41:06.839252 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" containerName="cluster-cloud-controller-manager" Dec 05 10:41:06.841055 master-0 kubenswrapper[10444]: I1205 10:41:06.840185 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:06.844902 master-0 kubenswrapper[10444]: I1205 10:41:06.842488 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"cloud-controller-manager-images" Dec 05 10:41:06.844902 master-0 kubenswrapper[10444]: I1205 10:41:06.844527 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:41:06.844902 master-0 kubenswrapper[10444]: I1205 10:41:06.844694 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"kube-rbac-proxy" Dec 05 10:41:06.844902 master-0 kubenswrapper[10444]: I1205 10:41:06.844823 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 10:41:06.848455 master-0 kubenswrapper[10444]: I1205 10:41:06.848066 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-controller-manager-operator"/"cluster-cloud-controller-manager-dockercfg-45kcs" Dec 05 10:41:06.848455 master-0 kubenswrapper[10444]: I1205 10:41:06.848320 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-controller-manager-operator"/"cloud-controller-manager-operator-tls" Dec 05 10:41:06.868404 master-0 kubenswrapper[10444]: I1205 10:41:06.868360 10444 scope.go:117] "RemoveContainer" containerID="9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c" Dec 05 10:41:06.868965 master-0 kubenswrapper[10444]: E1205 10:41:06.868912 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c\": container with ID starting with 9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c not found: ID does not exist" containerID="9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c" Dec 05 10:41:06.869084 master-0 kubenswrapper[10444]: I1205 10:41:06.869050 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c"} err="failed to get container status \"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c\": rpc error: code = NotFound desc = could not find container \"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c\": container with ID starting with 9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c not found: ID does not exist" Dec 05 10:41:06.869171 master-0 kubenswrapper[10444]: I1205 10:41:06.869086 10444 scope.go:117] "RemoveContainer" containerID="b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83" Dec 05 10:41:06.870279 master-0 kubenswrapper[10444]: E1205 10:41:06.870247 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83\": container with ID starting with b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83 not found: ID does not exist" containerID="b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83" Dec 05 10:41:06.870437 master-0 kubenswrapper[10444]: I1205 10:41:06.870285 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83"} err="failed to get container status \"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83\": rpc error: code = NotFound desc = could not find container \"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83\": container with ID starting with b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83 not found: ID does not exist" Dec 05 10:41:06.870437 master-0 kubenswrapper[10444]: I1205 10:41:06.870308 10444 scope.go:117] "RemoveContainer" containerID="29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05" Dec 05 10:41:06.876715 master-0 kubenswrapper[10444]: E1205 10:41:06.876663 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05\": container with ID starting with 29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05 not found: ID does not exist" containerID="29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05" Dec 05 10:41:06.876833 master-0 kubenswrapper[10444]: I1205 10:41:06.876711 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05"} err="failed to get container status \"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05\": rpc error: code = NotFound desc = could not find container \"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05\": container with ID starting with 29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05 not found: ID does not exist" Dec 05 10:41:06.876833 master-0 kubenswrapper[10444]: I1205 10:41:06.876740 10444 scope.go:117] "RemoveContainer" containerID="9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c" Dec 05 10:41:06.877454 master-0 kubenswrapper[10444]: I1205 10:41:06.877381 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c"} err="failed to get container status \"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c\": rpc error: code = NotFound desc = could not find container \"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c\": container with ID starting with 9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c not found: ID does not exist" Dec 05 10:41:06.877454 master-0 kubenswrapper[10444]: I1205 10:41:06.877452 10444 scope.go:117] "RemoveContainer" containerID="b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83" Dec 05 10:41:06.877982 master-0 kubenswrapper[10444]: I1205 10:41:06.877934 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83"} err="failed to get container status \"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83\": rpc error: code = NotFound desc = could not find container \"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83\": container with ID starting with b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83 not found: ID does not exist" Dec 05 10:41:06.877982 master-0 kubenswrapper[10444]: I1205 10:41:06.877958 10444 scope.go:117] "RemoveContainer" containerID="29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05" Dec 05 10:41:06.878460 master-0 kubenswrapper[10444]: I1205 10:41:06.878379 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05"} err="failed to get container status \"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05\": rpc error: code = NotFound desc = could not find container \"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05\": container with ID starting with 29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05 not found: ID does not exist" Dec 05 10:41:06.878460 master-0 kubenswrapper[10444]: I1205 10:41:06.878457 10444 scope.go:117] "RemoveContainer" containerID="9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c" Dec 05 10:41:06.878891 master-0 kubenswrapper[10444]: I1205 10:41:06.878801 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c"} err="failed to get container status \"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c\": rpc error: code = NotFound desc = could not find container \"9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c\": container with ID starting with 9b3dc6403c48451af8709221f0f8aee16c8a17143526c75c8ddc607f02b7745c not found: ID does not exist" Dec 05 10:41:06.878891 master-0 kubenswrapper[10444]: I1205 10:41:06.878873 10444 scope.go:117] "RemoveContainer" containerID="b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83" Dec 05 10:41:06.879581 master-0 kubenswrapper[10444]: I1205 10:41:06.879471 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83"} err="failed to get container status \"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83\": rpc error: code = NotFound desc = could not find container \"b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83\": container with ID starting with b2f7cb2e68dae400654ebf2d0d1b5e516838686ec69531fe5dfa6d325af86e83 not found: ID does not exist" Dec 05 10:41:06.879581 master-0 kubenswrapper[10444]: I1205 10:41:06.879494 10444 scope.go:117] "RemoveContainer" containerID="29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05" Dec 05 10:41:06.879727 master-0 kubenswrapper[10444]: I1205 10:41:06.879676 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05"} err="failed to get container status \"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05\": rpc error: code = NotFound desc = could not find container \"29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05\": container with ID starting with 29bbd41b280f57be37ed30fae778a38a6d4e4f33fbf38bbc08754f11b1808f05 not found: ID does not exist" Dec 05 10:41:06.962244 master-0 kubenswrapper[10444]: I1205 10:41:06.962164 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljdt5\" (UniqueName: \"kubernetes.io/projected/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-kube-api-access-ljdt5\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:06.962244 master-0 kubenswrapper[10444]: I1205 10:41:06.962230 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-auth-proxy-config\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:06.962244 master-0 kubenswrapper[10444]: I1205 10:41:06.962253 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-host-etc-kube\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:06.962654 master-0 kubenswrapper[10444]: I1205 10:41:06.962282 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cloud-controller-manager-operator-tls\" (UniqueName: \"kubernetes.io/secret/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-cloud-controller-manager-operator-tls\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:06.962654 master-0 kubenswrapper[10444]: I1205 10:41:06.962314 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-images\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.063319 master-0 kubenswrapper[10444]: I1205 10:41:07.063263 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-auth-proxy-config\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.063319 master-0 kubenswrapper[10444]: I1205 10:41:07.063310 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-host-etc-kube\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.065714 master-0 kubenswrapper[10444]: I1205 10:41:07.063345 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloud-controller-manager-operator-tls\" (UniqueName: \"kubernetes.io/secret/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-cloud-controller-manager-operator-tls\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.065714 master-0 kubenswrapper[10444]: I1205 10:41:07.063623 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-images\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.065714 master-0 kubenswrapper[10444]: I1205 10:41:07.063946 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljdt5\" (UniqueName: \"kubernetes.io/projected/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-kube-api-access-ljdt5\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.065714 master-0 kubenswrapper[10444]: I1205 10:41:07.064217 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-auth-proxy-config\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.065714 master-0 kubenswrapper[10444]: I1205 10:41:07.064171 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-host-etc-kube\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.065714 master-0 kubenswrapper[10444]: I1205 10:41:07.064941 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-images\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.069844 master-0 kubenswrapper[10444]: I1205 10:41:07.069759 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloud-controller-manager-operator-tls\" (UniqueName: \"kubernetes.io/secret/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-cloud-controller-manager-operator-tls\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.082827 master-0 kubenswrapper[10444]: I1205 10:41:07.082776 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljdt5\" (UniqueName: \"kubernetes.io/projected/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-kube-api-access-ljdt5\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.164221 master-0 kubenswrapper[10444]: I1205 10:41:07.164140 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:41:07.188132 master-0 kubenswrapper[10444]: W1205 10:41:07.188066 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf7078c2_ad4f_415b_ba2f_77f5145c3b3f.slice/crio-b0fb01570845c0cbcb28b2232b0ade7593b0c0de030669cae5f3ef53544cda81 WatchSource:0}: Error finding container b0fb01570845c0cbcb28b2232b0ade7593b0c0de030669cae5f3ef53544cda81: Status 404 returned error can't find the container with id b0fb01570845c0cbcb28b2232b0ade7593b0c0de030669cae5f3ef53544cda81 Dec 05 10:41:07.607641 master-0 kubenswrapper[10444]: I1205 10:41:07.607497 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a73d181f-f4f8-496a-a6f3-8e10804b394f" path="/var/lib/kubelet/pods/a73d181f-f4f8-496a-a6f3-8e10804b394f/volumes" Dec 05 10:41:07.751701 master-0 kubenswrapper[10444]: I1205 10:41:07.751617 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" event={"ID":"af7078c2-ad4f-415b-ba2f-77f5145c3b3f","Type":"ContainerStarted","Data":"76bf410edb844c728d030254090fc2f3989529da434d1e83243542e0c65e53f7"} Dec 05 10:41:07.751701 master-0 kubenswrapper[10444]: I1205 10:41:07.751674 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" event={"ID":"af7078c2-ad4f-415b-ba2f-77f5145c3b3f","Type":"ContainerStarted","Data":"b0fb01570845c0cbcb28b2232b0ade7593b0c0de030669cae5f3ef53544cda81"} Dec 05 10:41:07.790030 master-0 kubenswrapper[10444]: I1205 10:41:07.789935 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs"] Dec 05 10:41:07.790998 master-0 kubenswrapper[10444]: I1205 10:41:07.790826 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:41:07.796202 master-0 kubenswrapper[10444]: I1205 10:41:07.796143 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-brtjg" Dec 05 10:41:07.799145 master-0 kubenswrapper[10444]: I1205 10:41:07.798136 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 10:41:07.805571 master-0 kubenswrapper[10444]: I1205 10:41:07.805525 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs"] Dec 05 10:41:07.977465 master-0 kubenswrapper[10444]: I1205 10:41:07.977402 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-784mb\" (UniqueName: \"kubernetes.io/projected/6728a458-098b-44f9-8c8f-b0d76c5825fa-kube-api-access-784mb\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:41:07.977660 master-0 kubenswrapper[10444]: I1205 10:41:07.977565 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6728a458-098b-44f9-8c8f-b0d76c5825fa-mcc-auth-proxy-config\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:41:07.977721 master-0 kubenswrapper[10444]: I1205 10:41:07.977653 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6728a458-098b-44f9-8c8f-b0d76c5825fa-proxy-tls\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:41:08.078471 master-0 kubenswrapper[10444]: I1205 10:41:08.078405 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6728a458-098b-44f9-8c8f-b0d76c5825fa-proxy-tls\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:41:08.078695 master-0 kubenswrapper[10444]: I1205 10:41:08.078552 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-784mb\" (UniqueName: \"kubernetes.io/projected/6728a458-098b-44f9-8c8f-b0d76c5825fa-kube-api-access-784mb\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:41:08.078695 master-0 kubenswrapper[10444]: I1205 10:41:08.078589 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6728a458-098b-44f9-8c8f-b0d76c5825fa-mcc-auth-proxy-config\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:41:08.079801 master-0 kubenswrapper[10444]: I1205 10:41:08.079761 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6728a458-098b-44f9-8c8f-b0d76c5825fa-mcc-auth-proxy-config\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:41:08.084739 master-0 kubenswrapper[10444]: I1205 10:41:08.084544 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6728a458-098b-44f9-8c8f-b0d76c5825fa-proxy-tls\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:41:08.097656 master-0 kubenswrapper[10444]: I1205 10:41:08.097622 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-784mb\" (UniqueName: \"kubernetes.io/projected/6728a458-098b-44f9-8c8f-b0d76c5825fa-kube-api-access-784mb\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:41:08.122373 master-0 kubenswrapper[10444]: I1205 10:41:08.122331 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:41:08.530352 master-0 kubenswrapper[10444]: I1205 10:41:08.530311 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs"] Dec 05 10:41:08.761574 master-0 kubenswrapper[10444]: I1205 10:41:08.761523 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" event={"ID":"af7078c2-ad4f-415b-ba2f-77f5145c3b3f","Type":"ContainerStarted","Data":"6ba7074f6bb0f374dabeac1deb54c10aca607e1ffaef2825067cda94e2a5c8fd"} Dec 05 10:41:08.762026 master-0 kubenswrapper[10444]: I1205 10:41:08.761587 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" event={"ID":"af7078c2-ad4f-415b-ba2f-77f5145c3b3f","Type":"ContainerStarted","Data":"db1dbdb77596da747f1d050879a06169468c0d87025eceb17a18f84d5df289c8"} Dec 05 10:41:08.764587 master-0 kubenswrapper[10444]: I1205 10:41:08.764332 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" event={"ID":"6728a458-098b-44f9-8c8f-b0d76c5825fa","Type":"ContainerStarted","Data":"be0d1689f58c179873ee2f77d376672feba6d822713ae9575676444400388062"} Dec 05 10:41:08.764587 master-0 kubenswrapper[10444]: I1205 10:41:08.764370 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" event={"ID":"6728a458-098b-44f9-8c8f-b0d76c5825fa","Type":"ContainerStarted","Data":"45e53c8beb8eb2f205541b531720f581844942fc2ea4f6200401dc7c6d748d82"} Dec 05 10:41:08.788871 master-0 kubenswrapper[10444]: I1205 10:41:08.788798 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" podStartSLOduration=2.788781437 podStartE2EDuration="2.788781437s" podCreationTimestamp="2025-12-05 10:41:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:41:08.782630898 +0000 UTC m=+209.374543465" watchObservedRunningTime="2025-12-05 10:41:08.788781437 +0000 UTC m=+209.380694004" Dec 05 10:41:08.883365 master-0 kubenswrapper[10444]: I1205 10:41:08.883301 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw"] Dec 05 10:41:08.884248 master-0 kubenswrapper[10444]: I1205 10:41:08.884220 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" Dec 05 10:41:08.890285 master-0 kubenswrapper[10444]: I1205 10:41:08.890193 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5465c8b4db-s4c2f"] Dec 05 10:41:08.892499 master-0 kubenswrapper[10444]: I1205 10:41:08.891152 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:08.892790 master-0 kubenswrapper[10444]: I1205 10:41:08.892651 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 10:41:08.892790 master-0 kubenswrapper[10444]: I1205 10:41:08.892706 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 10:41:08.893160 master-0 kubenswrapper[10444]: I1205 10:41:08.893137 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 10:41:08.893762 master-0 kubenswrapper[10444]: I1205 10:41:08.893565 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 10:41:08.893762 master-0 kubenswrapper[10444]: I1205 10:41:08.893675 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 10:41:08.893966 master-0 kubenswrapper[10444]: I1205 10:41:08.893953 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 10:41:08.897999 master-0 kubenswrapper[10444]: I1205 10:41:08.897961 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz"] Dec 05 10:41:08.898768 master-0 kubenswrapper[10444]: I1205 10:41:08.898741 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" Dec 05 10:41:08.900489 master-0 kubenswrapper[10444]: I1205 10:41:08.900458 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Dec 05 10:41:08.900546 master-0 kubenswrapper[10444]: I1205 10:41:08.900501 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw"] Dec 05 10:41:08.904005 master-0 kubenswrapper[10444]: I1205 10:41:08.903974 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz"] Dec 05 10:41:08.990243 master-0 kubenswrapper[10444]: I1205 10:41:08.990194 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7824l\" (UniqueName: \"kubernetes.io/projected/da9d48c9-6346-4c9f-a690-f7419499c3e6-kube-api-access-7824l\") pod \"network-check-source-85d8db45d4-c2mhw\" (UID: \"da9d48c9-6346-4c9f-a690-f7419499c3e6\") " pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" Dec 05 10:41:08.990482 master-0 kubenswrapper[10444]: I1205 10:41:08.990287 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/0babac16-164a-405a-a86c-30524118f046-tls-certificates\") pod \"prometheus-operator-admission-webhook-7c85c4dffd-vjvbz\" (UID: \"0babac16-164a-405a-a86c-30524118f046\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" Dec 05 10:41:08.990482 master-0 kubenswrapper[10444]: I1205 10:41:08.990327 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-default-certificate\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:08.990482 master-0 kubenswrapper[10444]: I1205 10:41:08.990355 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-stats-auth\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:08.990623 master-0 kubenswrapper[10444]: I1205 10:41:08.990477 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-metrics-certs\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:08.990623 master-0 kubenswrapper[10444]: I1205 10:41:08.990546 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/445d75af-d072-4fa0-91a7-f3fa579b9ca9-service-ca-bundle\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:08.990700 master-0 kubenswrapper[10444]: I1205 10:41:08.990658 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hskm2\" (UniqueName: \"kubernetes.io/projected/445d75af-d072-4fa0-91a7-f3fa579b9ca9-kube-api-access-hskm2\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.091564 master-0 kubenswrapper[10444]: I1205 10:41:09.091434 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-default-certificate\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.091564 master-0 kubenswrapper[10444]: I1205 10:41:09.091515 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-stats-auth\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.091564 master-0 kubenswrapper[10444]: I1205 10:41:09.091561 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-metrics-certs\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.091805 master-0 kubenswrapper[10444]: I1205 10:41:09.091591 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/445d75af-d072-4fa0-91a7-f3fa579b9ca9-service-ca-bundle\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.091805 master-0 kubenswrapper[10444]: I1205 10:41:09.091639 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hskm2\" (UniqueName: \"kubernetes.io/projected/445d75af-d072-4fa0-91a7-f3fa579b9ca9-kube-api-access-hskm2\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.091871 master-0 kubenswrapper[10444]: I1205 10:41:09.091795 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7824l\" (UniqueName: \"kubernetes.io/projected/da9d48c9-6346-4c9f-a690-f7419499c3e6-kube-api-access-7824l\") pod \"network-check-source-85d8db45d4-c2mhw\" (UID: \"da9d48c9-6346-4c9f-a690-f7419499c3e6\") " pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" Dec 05 10:41:09.091871 master-0 kubenswrapper[10444]: I1205 10:41:09.091842 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/0babac16-164a-405a-a86c-30524118f046-tls-certificates\") pod \"prometheus-operator-admission-webhook-7c85c4dffd-vjvbz\" (UID: \"0babac16-164a-405a-a86c-30524118f046\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" Dec 05 10:41:09.093531 master-0 kubenswrapper[10444]: I1205 10:41:09.093370 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/445d75af-d072-4fa0-91a7-f3fa579b9ca9-service-ca-bundle\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.094743 master-0 kubenswrapper[10444]: I1205 10:41:09.094705 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-default-certificate\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.096893 master-0 kubenswrapper[10444]: I1205 10:41:09.096860 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-metrics-certs\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.098600 master-0 kubenswrapper[10444]: I1205 10:41:09.098572 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/0babac16-164a-405a-a86c-30524118f046-tls-certificates\") pod \"prometheus-operator-admission-webhook-7c85c4dffd-vjvbz\" (UID: \"0babac16-164a-405a-a86c-30524118f046\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" Dec 05 10:41:09.099410 master-0 kubenswrapper[10444]: I1205 10:41:09.099376 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-stats-auth\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.275893 master-0 kubenswrapper[10444]: I1205 10:41:09.275819 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" Dec 05 10:41:09.293318 master-0 kubenswrapper[10444]: I1205 10:41:09.293241 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hskm2\" (UniqueName: \"kubernetes.io/projected/445d75af-d072-4fa0-91a7-f3fa579b9ca9-kube-api-access-hskm2\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.299919 master-0 kubenswrapper[10444]: I1205 10:41:09.299886 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7824l\" (UniqueName: \"kubernetes.io/projected/da9d48c9-6346-4c9f-a690-f7419499c3e6-kube-api-access-7824l\") pod \"network-check-source-85d8db45d4-c2mhw\" (UID: \"da9d48c9-6346-4c9f-a690-f7419499c3e6\") " pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" Dec 05 10:41:09.300219 master-0 kubenswrapper[10444]: I1205 10:41:09.300170 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7824l\" (UniqueName: \"kubernetes.io/projected/da9d48c9-6346-4c9f-a690-f7419499c3e6-kube-api-access-7824l\") pod \"network-check-source-85d8db45d4-c2mhw\" (UID: \"da9d48c9-6346-4c9f-a690-f7419499c3e6\") " pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" Dec 05 10:41:09.301160 master-0 kubenswrapper[10444]: I1205 10:41:09.301140 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7824l\" (UniqueName: \"kubernetes.io/projected/da9d48c9-6346-4c9f-a690-f7419499c3e6-kube-api-access-7824l\") pod \"network-check-source-85d8db45d4-c2mhw\" (UID: \"da9d48c9-6346-4c9f-a690-f7419499c3e6\") " pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" Dec 05 10:41:09.537328 master-0 kubenswrapper[10444]: I1205 10:41:09.536792 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" Dec 05 10:41:09.555931 master-0 kubenswrapper[10444]: I1205 10:41:09.555868 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:09.589369 master-0 kubenswrapper[10444]: W1205 10:41:09.589293 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod445d75af_d072_4fa0_91a7_f3fa579b9ca9.slice/crio-5cd120c4a52499689b34a1d9baed806acbb93f12cd8a2bcbf5cdb4f6c23214df WatchSource:0}: Error finding container 5cd120c4a52499689b34a1d9baed806acbb93f12cd8a2bcbf5cdb4f6c23214df: Status 404 returned error can't find the container with id 5cd120c4a52499689b34a1d9baed806acbb93f12cd8a2bcbf5cdb4f6c23214df Dec 05 10:41:09.717259 master-0 kubenswrapper[10444]: I1205 10:41:09.717101 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz"] Dec 05 10:41:09.722131 master-0 kubenswrapper[10444]: W1205 10:41:09.722080 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0babac16_164a_405a_a86c_30524118f046.slice/crio-812553092d13019021bace1ffb05fbafd76c83d356691f1b938f38559234f787 WatchSource:0}: Error finding container 812553092d13019021bace1ffb05fbafd76c83d356691f1b938f38559234f787: Status 404 returned error can't find the container with id 812553092d13019021bace1ffb05fbafd76c83d356691f1b938f38559234f787 Dec 05 10:41:09.777532 master-0 kubenswrapper[10444]: I1205 10:41:09.777471 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" event={"ID":"6728a458-098b-44f9-8c8f-b0d76c5825fa","Type":"ContainerStarted","Data":"baf0d4ad553c4845e2843145f59d20d71f0982328f647df92245559bf6c037dd"} Dec 05 10:41:09.779353 master-0 kubenswrapper[10444]: I1205 10:41:09.779302 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" event={"ID":"445d75af-d072-4fa0-91a7-f3fa579b9ca9","Type":"ContainerStarted","Data":"5cd120c4a52499689b34a1d9baed806acbb93f12cd8a2bcbf5cdb4f6c23214df"} Dec 05 10:41:09.781543 master-0 kubenswrapper[10444]: I1205 10:41:09.781501 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" event={"ID":"0babac16-164a-405a-a86c-30524118f046","Type":"ContainerStarted","Data":"812553092d13019021bace1ffb05fbafd76c83d356691f1b938f38559234f787"} Dec 05 10:41:09.804049 master-0 kubenswrapper[10444]: I1205 10:41:09.803961 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" podStartSLOduration=2.803935696 podStartE2EDuration="2.803935696s" podCreationTimestamp="2025-12-05 10:41:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:41:09.797268624 +0000 UTC m=+210.389181191" watchObservedRunningTime="2025-12-05 10:41:09.803935696 +0000 UTC m=+210.395848263" Dec 05 10:41:10.004727 master-0 kubenswrapper[10444]: I1205 10:41:10.004609 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw"] Dec 05 10:41:10.009531 master-0 kubenswrapper[10444]: W1205 10:41:10.009499 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda9d48c9_6346_4c9f_a690_f7419499c3e6.slice/crio-61b925377e25c3f9cf64e2a949ea1fe86f2ea5c4e951a4e7adc2aa3b0cb3336d WatchSource:0}: Error finding container 61b925377e25c3f9cf64e2a949ea1fe86f2ea5c4e951a4e7adc2aa3b0cb3336d: Status 404 returned error can't find the container with id 61b925377e25c3f9cf64e2a949ea1fe86f2ea5c4e951a4e7adc2aa3b0cb3336d Dec 05 10:41:10.788864 master-0 kubenswrapper[10444]: I1205 10:41:10.788769 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" event={"ID":"da9d48c9-6346-4c9f-a690-f7419499c3e6","Type":"ContainerStarted","Data":"c52e04e4c01fb4fb23adfe05c796b9b28ca823366e0cb818898f1ad7e433e11b"} Dec 05 10:41:10.788864 master-0 kubenswrapper[10444]: I1205 10:41:10.788842 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" event={"ID":"da9d48c9-6346-4c9f-a690-f7419499c3e6","Type":"ContainerStarted","Data":"61b925377e25c3f9cf64e2a949ea1fe86f2ea5c4e951a4e7adc2aa3b0cb3336d"} Dec 05 10:41:10.869348 master-0 kubenswrapper[10444]: I1205 10:41:10.869249 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" podStartSLOduration=277.869232736 podStartE2EDuration="4m37.869232736s" podCreationTimestamp="2025-12-05 10:36:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:41:10.86633795 +0000 UTC m=+211.458250517" watchObservedRunningTime="2025-12-05 10:41:10.869232736 +0000 UTC m=+211.461145303" Dec 05 10:41:11.019553 master-0 kubenswrapper[10444]: I1205 10:41:11.019504 10444 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 10:41:12.324053 master-0 kubenswrapper[10444]: I1205 10:41:12.323963 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8_d95a56ba-c940-4e3e-aed6-d8c04f1871b6/authentication-operator/0.log" Dec 05 10:41:12.517506 master-0 kubenswrapper[10444]: I1205 10:41:12.517457 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8_d95a56ba-c940-4e3e-aed6-d8c04f1871b6/authentication-operator/1.log" Dec 05 10:41:12.915182 master-0 kubenswrapper[10444]: I1205 10:41:12.915135 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-oauth-apiserver_apiserver-85b8f855df-8g52w_ca45c52e-fb30-4e7c-8c3f-e685c0909916/fix-audit-permissions/0.log" Dec 05 10:41:13.118644 master-0 kubenswrapper[10444]: I1205 10:41:13.118568 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-oauth-apiserver_apiserver-85b8f855df-8g52w_ca45c52e-fb30-4e7c-8c3f-e685c0909916/oauth-apiserver/0.log" Dec 05 10:41:13.237219 master-0 kubenswrapper[10444]: I1205 10:41:13.237002 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-5t4nn"] Dec 05 10:41:13.237951 master-0 kubenswrapper[10444]: I1205 10:41:13.237912 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:41:13.239794 master-0 kubenswrapper[10444]: I1205 10:41:13.239763 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 10:41:13.240031 master-0 kubenswrapper[10444]: I1205 10:41:13.240003 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 10:41:13.240189 master-0 kubenswrapper[10444]: I1205 10:41:13.240161 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-zlzd6" Dec 05 10:41:13.269341 master-0 kubenswrapper[10444]: I1205 10:41:13.269287 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-node-bootstrap-token\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:41:13.269537 master-0 kubenswrapper[10444]: I1205 10:41:13.269367 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-certs\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:41:13.269641 master-0 kubenswrapper[10444]: I1205 10:41:13.269590 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flvlq\" (UniqueName: \"kubernetes.io/projected/568e9bed-cbd2-49bd-84af-9e17f95a003f-kube-api-access-flvlq\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:41:13.322262 master-0 kubenswrapper[10444]: I1205 10:41:13.322192 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-autoscaler-operator-5f49d774cd-cfg5f_1f132702-304b-46ff-b428-6df1deeffec3/kube-rbac-proxy/0.log" Dec 05 10:41:13.370846 master-0 kubenswrapper[10444]: I1205 10:41:13.370759 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flvlq\" (UniqueName: \"kubernetes.io/projected/568e9bed-cbd2-49bd-84af-9e17f95a003f-kube-api-access-flvlq\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:41:13.370846 master-0 kubenswrapper[10444]: I1205 10:41:13.370826 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-node-bootstrap-token\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:41:13.371621 master-0 kubenswrapper[10444]: I1205 10:41:13.370871 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-certs\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:41:13.373887 master-0 kubenswrapper[10444]: I1205 10:41:13.373853 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-certs\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:41:13.391411 master-0 kubenswrapper[10444]: I1205 10:41:13.391358 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-node-bootstrap-token\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:41:13.392733 master-0 kubenswrapper[10444]: I1205 10:41:13.392682 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flvlq\" (UniqueName: \"kubernetes.io/projected/568e9bed-cbd2-49bd-84af-9e17f95a003f-kube-api-access-flvlq\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:41:13.519214 master-0 kubenswrapper[10444]: I1205 10:41:13.518756 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-autoscaler-operator-5f49d774cd-cfg5f_1f132702-304b-46ff-b428-6df1deeffec3/cluster-autoscaler-operator/0.log" Dec 05 10:41:13.562069 master-0 kubenswrapper[10444]: I1205 10:41:13.561993 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:41:13.718518 master-0 kubenswrapper[10444]: I1205 10:41:13.718458 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-baremetal-operator-78f758c7b9-6t2gm_48bd1d86-a6f2-439f-ab04-6a9a442bec42/cluster-baremetal-operator/0.log" Dec 05 10:41:14.070406 master-0 kubenswrapper[10444]: I1205 10:41:14.069537 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-baremetal-operator-78f758c7b9-6t2gm_48bd1d86-a6f2-439f-ab04-6a9a442bec42/baremetal-kube-rbac-proxy/0.log" Dec 05 10:41:14.114404 master-0 kubenswrapper[10444]: I1205 10:41:14.114328 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-7df95c79b5-qnq6t_24444474-aa9a-4a0f-8b4d-90f0009e0dc7/control-plane-machine-set-operator/0.log" Dec 05 10:41:14.323549 master-0 kubenswrapper[10444]: I1205 10:41:14.322509 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-88d48b57d-x7jfs_82464e92-4de2-48f0-8772-a489abb16898/kube-rbac-proxy/0.log" Dec 05 10:41:14.519202 master-0 kubenswrapper[10444]: I1205 10:41:14.519135 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-88d48b57d-x7jfs_82464e92-4de2-48f0-8772-a489abb16898/machine-api-operator/0.log" Dec 05 10:41:14.721721 master-0 kubenswrapper[10444]: I1205 10:41:14.721596 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c_f7a08359-0379-4364-8b0c-ddb58ff605f4/etcd-operator/0.log" Dec 05 10:41:14.916387 master-0 kubenswrapper[10444]: I1205 10:41:14.916315 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c_f7a08359-0379-4364-8b0c-ddb58ff605f4/etcd-operator/1.log" Dec 05 10:41:15.115645 master-0 kubenswrapper[10444]: I1205 10:41:15.115581 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/setup/0.log" Dec 05 10:41:15.315973 master-0 kubenswrapper[10444]: I1205 10:41:15.315914 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd-ensure-env-vars/0.log" Dec 05 10:41:15.516792 master-0 kubenswrapper[10444]: I1205 10:41:15.516685 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd-resources-copy/0.log" Dec 05 10:41:15.718140 master-0 kubenswrapper[10444]: I1205 10:41:15.716614 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcdctl/0.log" Dec 05 10:41:15.771475 master-0 kubenswrapper[10444]: I1205 10:41:15.771355 10444 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["kube-system/bootstrap-kube-controller-manager-master-0"] Dec 05 10:41:15.771596 master-0 kubenswrapper[10444]: I1205 10:41:15.771492 10444 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-master-0"] Dec 05 10:41:15.772048 master-0 kubenswrapper[10444]: I1205 10:41:15.771972 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="kube-system/bootstrap-kube-controller-manager-master-0" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="cluster-policy-controller" containerID="cri-o://80aee350277cce308d84133aa5eefc6be3ed0712557f82d0d22b78dac448f0af" gracePeriod=30 Dec 05 10:41:15.772562 master-0 kubenswrapper[10444]: E1205 10:41:15.772523 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="cluster-policy-controller" Dec 05 10:41:15.772562 master-0 kubenswrapper[10444]: I1205 10:41:15.772560 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="cluster-policy-controller" Dec 05 10:41:15.772641 master-0 kubenswrapper[10444]: E1205 10:41:15.772596 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.772641 master-0 kubenswrapper[10444]: I1205 10:41:15.772609 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.772641 master-0 kubenswrapper[10444]: I1205 10:41:15.772606 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="kube-system/bootstrap-kube-controller-manager-master-0" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" containerID="cri-o://efee17240223c7e06e121d89cbef5a60b4ccceeebdd17b05f12e6a98b82892f6" gracePeriod=30 Dec 05 10:41:15.772641 master-0 kubenswrapper[10444]: E1205 10:41:15.772642 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.772782 master-0 kubenswrapper[10444]: I1205 10:41:15.772655 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.772782 master-0 kubenswrapper[10444]: E1205 10:41:15.772677 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.772782 master-0 kubenswrapper[10444]: I1205 10:41:15.772690 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.772991 master-0 kubenswrapper[10444]: I1205 10:41:15.772961 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.773031 master-0 kubenswrapper[10444]: I1205 10:41:15.772988 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.773031 master-0 kubenswrapper[10444]: I1205 10:41:15.773006 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="cluster-policy-controller" Dec 05 10:41:15.773031 master-0 kubenswrapper[10444]: I1205 10:41:15.773023 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.773592 master-0 kubenswrapper[10444]: E1205 10:41:15.773192 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.773592 master-0 kubenswrapper[10444]: I1205 10:41:15.773209 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.773592 master-0 kubenswrapper[10444]: I1205 10:41:15.773361 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="kube-controller-manager" Dec 05 10:41:15.774456 master-0 kubenswrapper[10444]: I1205 10:41:15.774393 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:15.802282 master-0 kubenswrapper[10444]: I1205 10:41:15.802234 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"a51fa80258077cf0f6fddf343b4ba54f\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:15.802378 master-0 kubenswrapper[10444]: I1205 10:41:15.802295 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"a51fa80258077cf0f6fddf343b4ba54f\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:15.903286 master-0 kubenswrapper[10444]: I1205 10:41:15.903233 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"a51fa80258077cf0f6fddf343b4ba54f\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:15.903645 master-0 kubenswrapper[10444]: I1205 10:41:15.903374 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"a51fa80258077cf0f6fddf343b4ba54f\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:15.903645 master-0 kubenswrapper[10444]: I1205 10:41:15.903494 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"a51fa80258077cf0f6fddf343b4ba54f\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:15.903645 master-0 kubenswrapper[10444]: I1205 10:41:15.903615 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"a51fa80258077cf0f6fddf343b4ba54f\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:16.839268 master-0 kubenswrapper[10444]: I1205 10:41:16.839206 10444 generic.go:334] "Generic (PLEG): container finished" podID="8b47694fcc32464ab24d09c23d6efb57" containerID="efee17240223c7e06e121d89cbef5a60b4ccceeebdd17b05f12e6a98b82892f6" exitCode=0 Dec 05 10:41:16.840052 master-0 kubenswrapper[10444]: I1205 10:41:16.839293 10444 scope.go:117] "RemoveContainer" containerID="e7ce8dc7bd304b25a361ecabc9d8faa1f0ae1eca9ef1f9dc24bcebc4d09b4a20" Dec 05 10:41:17.538792 master-0 kubenswrapper[10444]: I1205 10:41:17.538692 10444 prober.go:107] "Probe failed" probeType="Readiness" pod="kube-system/bootstrap-kube-controller-manager-master-0" podUID="8b47694fcc32464ab24d09c23d6efb57" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://localhost:10357/healthz\": dial tcp [::1]:10357: connect: connection refused" Dec 05 10:41:18.854868 master-0 kubenswrapper[10444]: I1205 10:41:18.854800 10444 generic.go:334] "Generic (PLEG): container finished" podID="8b47694fcc32464ab24d09c23d6efb57" containerID="80aee350277cce308d84133aa5eefc6be3ed0712557f82d0d22b78dac448f0af" exitCode=0 Dec 05 10:41:19.864017 master-0 kubenswrapper[10444]: I1205 10:41:19.863903 10444 generic.go:334] "Generic (PLEG): container finished" podID="70fbf150-c1f7-4857-8b63-c566282e2526" containerID="2215577fb4531349eb7213f364526a3db417316312ccf7c18444e96dbaf859c1" exitCode=0 Dec 05 10:41:19.864017 master-0 kubenswrapper[10444]: I1205 10:41:19.863960 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-2-master-0" event={"ID":"70fbf150-c1f7-4857-8b63-c566282e2526","Type":"ContainerDied","Data":"2215577fb4531349eb7213f364526a3db417316312ccf7c18444e96dbaf859c1"} Dec 05 10:41:22.071466 master-0 kubenswrapper[10444]: I1205 10:41:22.071374 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:22.073024 master-0 kubenswrapper[10444]: I1205 10:41:22.072728 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-master-0"] Dec 05 10:41:30.928152 master-0 kubenswrapper[10444]: I1205 10:41:30.928115 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd/0.log" Dec 05 10:41:30.963924 master-0 kubenswrapper[10444]: I1205 10:41:30.963483 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd-metrics/0.log" Dec 05 10:41:30.980515 master-0 kubenswrapper[10444]: I1205 10:41:30.980485 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd-readyz/0.log" Dec 05 10:41:30.992469 master-0 kubenswrapper[10444]: I1205 10:41:30.992415 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd-rev/0.log" Dec 05 10:41:31.002484 master-0 kubenswrapper[10444]: I1205 10:41:31.002398 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-1-master-0_e149cbe7-1ec3-4a06-af61-3a8906b8e9ef/installer/0.log" Dec 05 10:41:31.021940 master-0 kubenswrapper[10444]: I1205 10:41:31.021904 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl_444f8808-e454-4015-9e20-429e715a08c7/kube-apiserver-operator/0.log" Dec 05 10:41:31.046378 master-0 kubenswrapper[10444]: I1205 10:41:31.046348 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl_444f8808-e454-4015-9e20-429e715a08c7/kube-apiserver-operator/1.log" Dec 05 10:41:31.056531 master-0 kubenswrapper[10444]: I1205 10:41:31.056495 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_bootstrap-kube-apiserver-master-0_d75143d9bc4a2dc15781dc51ccff632a/setup/0.log" Dec 05 10:41:31.072923 master-0 kubenswrapper[10444]: I1205 10:41:31.072869 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_bootstrap-kube-apiserver-master-0_d75143d9bc4a2dc15781dc51ccff632a/kube-apiserver/0.log" Dec 05 10:41:31.112480 master-0 kubenswrapper[10444]: I1205 10:41:31.112389 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_bootstrap-kube-apiserver-master-0_d75143d9bc4a2dc15781dc51ccff632a/kube-apiserver-insecure-readyz/0.log" Dec 05 10:41:31.130489 master-0 kubenswrapper[10444]: I1205 10:41:31.130408 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-1-master-0_5d8c854f-31f8-42d3-b28c-82c7d1b14ee9/installer/0.log" Dec 05 10:41:31.142010 master-0 kubenswrapper[10444]: I1205 10:41:31.141943 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_installer-1-master-0_738d9fe7-2d08-4aa4-842b-9cbeddf120fb/installer/0.log" Dec 05 10:41:31.338127 master-0 kubenswrapper[10444]: I1205 10:41:31.338077 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_installer-2-master-0_70fbf150-c1f7-4857-8b63-c566282e2526/installer/0.log" Dec 05 10:41:31.540029 master-0 kubenswrapper[10444]: I1205 10:41:31.539971 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager-operator_kube-controller-manager-operator-848f645654-rmdb8_11f563d5-89bb-433c-956a-6d5d2492e8f1/kube-controller-manager-operator/0.log" Dec 05 10:41:31.740531 master-0 kubenswrapper[10444]: I1205 10:41:31.740076 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager-operator_kube-controller-manager-operator-848f645654-rmdb8_11f563d5-89bb-433c-956a-6d5d2492e8f1/kube-controller-manager-operator/1.log" Dec 05 10:41:32.739292 master-0 kubenswrapper[10444]: I1205 10:41:32.739255 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_bootstrap-kube-scheduler-master-0_5e09e2af7200e6f9be469dbfd9bb1127/kube-scheduler/0.log" Dec 05 10:41:35.342448 master-0 kubenswrapper[10444]: I1205 10:41:35.338894 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/kube-system_bootstrap-kube-scheduler-master-0_5e09e2af7200e6f9be469dbfd9bb1127/kube-scheduler/1.log" Dec 05 10:41:35.355625 master-0 kubenswrapper[10444]: I1205 10:41:35.355563 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler_installer-3-master-0_6325f9a2-86d5-4fac-9e71-331b2628f965/installer/0.log" Dec 05 10:41:36.812393 master-0 kubenswrapper[10444]: I1205 10:41:36.812302 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler-operator_openshift-kube-scheduler-operator-5f85974995-dwh5t_4825316a-ea9f-4d3d-838b-fa809a6e49c7/kube-scheduler-operator-container/0.log" Dec 05 10:41:36.829591 master-0 kubenswrapper[10444]: I1205 10:41:36.829508 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler-operator_openshift-kube-scheduler-operator-5f85974995-dwh5t_4825316a-ea9f-4d3d-838b-fa809a6e49c7/kube-scheduler-operator-container/1.log" Dec 05 10:41:36.841654 master-0 kubenswrapper[10444]: I1205 10:41:36.841582 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-apiserver-operator_openshift-apiserver-operator-7bf7f6b755-hdjv7_6f76d12f-5406-47e2-8337-2f50e35376d6/openshift-apiserver-operator/0.log" Dec 05 10:41:36.849321 master-0 kubenswrapper[10444]: I1205 10:41:36.849264 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-apiserver-operator_openshift-apiserver-operator-7bf7f6b755-hdjv7_6f76d12f-5406-47e2-8337-2f50e35376d6/openshift-apiserver-operator/1.log" Dec 05 10:41:36.855104 master-0 kubenswrapper[10444]: I1205 10:41:36.855050 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-apiserver_apiserver-5b9fd577f8-6sxcx_d5198438-06ae-4e63-a7e3-950ba23bba9c/fix-audit-permissions/0.log" Dec 05 10:41:36.864337 master-0 kubenswrapper[10444]: I1205 10:41:36.864291 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-apiserver_apiserver-5b9fd577f8-6sxcx_d5198438-06ae-4e63-a7e3-950ba23bba9c/openshift-apiserver/0.log" Dec 05 10:41:36.873467 master-0 kubenswrapper[10444]: I1205 10:41:36.873386 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-apiserver_apiserver-5b9fd577f8-6sxcx_d5198438-06ae-4e63-a7e3-950ba23bba9c/openshift-apiserver-check-endpoints/0.log" Dec 05 10:41:36.882684 master-0 kubenswrapper[10444]: I1205 10:41:36.882624 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c_f7a08359-0379-4364-8b0c-ddb58ff605f4/etcd-operator/0.log" Dec 05 10:41:36.889527 master-0 kubenswrapper[10444]: I1205 10:41:36.888479 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c_f7a08359-0379-4364-8b0c-ddb58ff605f4/etcd-operator/1.log" Dec 05 10:41:36.899565 master-0 kubenswrapper[10444]: I1205 10:41:36.899506 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_catalog-operator-fbc6455c4-mbm77_4c2975ec-e33d-4960-a708-277d41c79b15/catalog-operator/0.log" Dec 05 10:41:36.908587 master-0 kubenswrapper[10444]: I1205 10:41:36.908528 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_olm-operator-7cd7dbb44c-d25sk_46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/olm-operator/0.log" Dec 05 10:41:37.020862 master-0 kubenswrapper[10444]: I1205 10:41:37.020810 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_package-server-manager-67477646d4-nm8cn_8c649a16-c187-412e-b5da-62a00bee38ab/kube-rbac-proxy/0.log" Dec 05 10:41:37.224829 master-0 kubenswrapper[10444]: I1205 10:41:37.224685 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_package-server-manager-67477646d4-nm8cn_8c649a16-c187-412e-b5da-62a00bee38ab/package-server-manager/0.log" Dec 05 10:41:37.423444 master-0 kubenswrapper[10444]: I1205 10:41:37.423372 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-lifecycle-manager_packageserver-d7b67d8cf-krp6c_2c3fab4d-05b8-42ec-8c91-91ff64562649/packageserver/0.log" Dec 05 10:41:38.080926 master-0 kubenswrapper[10444]: I1205 10:41:38.080873 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:41:38.121541 master-0 kubenswrapper[10444]: W1205 10:41:38.121496 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod568e9bed_cbd2_49bd_84af_9e17f95a003f.slice/crio-d22a14317ab6ff7dd171c6dc6d6a0e336ee5087a9085305232d6accf32502f02 WatchSource:0}: Error finding container d22a14317ab6ff7dd171c6dc6d6a0e336ee5087a9085305232d6accf32502f02: Status 404 returned error can't find the container with id d22a14317ab6ff7dd171c6dc6d6a0e336ee5087a9085305232d6accf32502f02 Dec 05 10:41:38.132533 master-0 kubenswrapper[10444]: W1205 10:41:38.132481 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda51fa80258077cf0f6fddf343b4ba54f.slice/crio-512588b6c638c1ce7fb1dce16bbe4f74e94e57511137324c00c995274f332c5c WatchSource:0}: Error finding container 512588b6c638c1ce7fb1dce16bbe4f74e94e57511137324c00c995274f332c5c: Status 404 returned error can't find the container with id 512588b6c638c1ce7fb1dce16bbe4f74e94e57511137324c00c995274f332c5c Dec 05 10:41:38.210671 master-0 kubenswrapper[10444]: I1205 10:41:38.210620 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70fbf150-c1f7-4857-8b63-c566282e2526-kube-api-access\") pod \"70fbf150-c1f7-4857-8b63-c566282e2526\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " Dec 05 10:41:38.210795 master-0 kubenswrapper[10444]: I1205 10:41:38.210712 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-var-lock\") pod \"70fbf150-c1f7-4857-8b63-c566282e2526\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " Dec 05 10:41:38.210795 master-0 kubenswrapper[10444]: I1205 10:41:38.210774 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-kubelet-dir\") pod \"70fbf150-c1f7-4857-8b63-c566282e2526\" (UID: \"70fbf150-c1f7-4857-8b63-c566282e2526\") " Dec 05 10:41:38.210887 master-0 kubenswrapper[10444]: I1205 10:41:38.210843 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-var-lock" (OuterVolumeSpecName: "var-lock") pod "70fbf150-c1f7-4857-8b63-c566282e2526" (UID: "70fbf150-c1f7-4857-8b63-c566282e2526"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:41:38.210933 master-0 kubenswrapper[10444]: I1205 10:41:38.210911 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "70fbf150-c1f7-4857-8b63-c566282e2526" (UID: "70fbf150-c1f7-4857-8b63-c566282e2526"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:41:38.211182 master-0 kubenswrapper[10444]: I1205 10:41:38.211151 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:38.211242 master-0 kubenswrapper[10444]: I1205 10:41:38.211200 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/70fbf150-c1f7-4857-8b63-c566282e2526-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:38.215414 master-0 kubenswrapper[10444]: I1205 10:41:38.215367 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70fbf150-c1f7-4857-8b63-c566282e2526-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "70fbf150-c1f7-4857-8b63-c566282e2526" (UID: "70fbf150-c1f7-4857-8b63-c566282e2526"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:41:38.247799 master-0 kubenswrapper[10444]: I1205 10:41:38.247754 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:41:38.311746 master-0 kubenswrapper[10444]: I1205 10:41:38.311633 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-logs\") pod \"8b47694fcc32464ab24d09c23d6efb57\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " Dec 05 10:41:38.311746 master-0 kubenswrapper[10444]: I1205 10:41:38.311686 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-secrets\") pod \"8b47694fcc32464ab24d09c23d6efb57\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " Dec 05 10:41:38.311746 master-0 kubenswrapper[10444]: I1205 10:41:38.311709 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-etc-kubernetes-cloud\") pod \"8b47694fcc32464ab24d09c23d6efb57\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " Dec 05 10:41:38.311746 master-0 kubenswrapper[10444]: I1205 10:41:38.311732 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-ssl-certs-host\") pod \"8b47694fcc32464ab24d09c23d6efb57\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " Dec 05 10:41:38.311746 master-0 kubenswrapper[10444]: I1205 10:41:38.311756 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-config\") pod \"8b47694fcc32464ab24d09c23d6efb57\" (UID: \"8b47694fcc32464ab24d09c23d6efb57\") " Dec 05 10:41:38.311746 master-0 kubenswrapper[10444]: I1205 10:41:38.311763 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-logs" (OuterVolumeSpecName: "logs") pod "8b47694fcc32464ab24d09c23d6efb57" (UID: "8b47694fcc32464ab24d09c23d6efb57"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:41:38.312406 master-0 kubenswrapper[10444]: I1205 10:41:38.311807 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-secrets" (OuterVolumeSpecName: "secrets") pod "8b47694fcc32464ab24d09c23d6efb57" (UID: "8b47694fcc32464ab24d09c23d6efb57"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:41:38.312406 master-0 kubenswrapper[10444]: I1205 10:41:38.311847 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-ssl-certs-host" (OuterVolumeSpecName: "ssl-certs-host") pod "8b47694fcc32464ab24d09c23d6efb57" (UID: "8b47694fcc32464ab24d09c23d6efb57"). InnerVolumeSpecName "ssl-certs-host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:41:38.312406 master-0 kubenswrapper[10444]: I1205 10:41:38.311876 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-config" (OuterVolumeSpecName: "config") pod "8b47694fcc32464ab24d09c23d6efb57" (UID: "8b47694fcc32464ab24d09c23d6efb57"). InnerVolumeSpecName "config". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:41:38.312406 master-0 kubenswrapper[10444]: I1205 10:41:38.311864 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-etc-kubernetes-cloud" (OuterVolumeSpecName: "etc-kubernetes-cloud") pod "8b47694fcc32464ab24d09c23d6efb57" (UID: "8b47694fcc32464ab24d09c23d6efb57"). InnerVolumeSpecName "etc-kubernetes-cloud". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:41:38.312406 master-0 kubenswrapper[10444]: I1205 10:41:38.312123 10444 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-secrets\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:38.312406 master-0 kubenswrapper[10444]: I1205 10:41:38.312138 10444 reconciler_common.go:293] "Volume detached for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-etc-kubernetes-cloud\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:38.312406 master-0 kubenswrapper[10444]: I1205 10:41:38.312182 10444 reconciler_common.go:293] "Volume detached for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-ssl-certs-host\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:38.312406 master-0 kubenswrapper[10444]: I1205 10:41:38.312192 10444 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:38.312406 master-0 kubenswrapper[10444]: I1205 10:41:38.312201 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/70fbf150-c1f7-4857-8b63-c566282e2526-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:38.312406 master-0 kubenswrapper[10444]: I1205 10:41:38.312212 10444 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/8b47694fcc32464ab24d09c23d6efb57-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 10:41:38.985938 master-0 kubenswrapper[10444]: I1205 10:41:38.985856 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5t4nn" event={"ID":"568e9bed-cbd2-49bd-84af-9e17f95a003f","Type":"ContainerStarted","Data":"2ce6af3669bcb8adcc1ef51fdc02e005a5230862a290a874ff4489d6d01437d2"} Dec 05 10:41:38.985938 master-0 kubenswrapper[10444]: I1205 10:41:38.985924 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-5t4nn" event={"ID":"568e9bed-cbd2-49bd-84af-9e17f95a003f","Type":"ContainerStarted","Data":"d22a14317ab6ff7dd171c6dc6d6a0e336ee5087a9085305232d6accf32502f02"} Dec 05 10:41:38.993489 master-0 kubenswrapper[10444]: I1205 10:41:38.993443 10444 generic.go:334] "Generic (PLEG): container finished" podID="489b6771-aa16-469f-9883-9a9233d3c379" containerID="0a06bffef31dc5f888e335796830f40b7d25388d95911101d91480b1460e1c2e" exitCode=0 Dec 05 10:41:38.993845 master-0 kubenswrapper[10444]: I1205 10:41:38.993660 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djhk8" event={"ID":"489b6771-aa16-469f-9883-9a9233d3c379","Type":"ContainerDied","Data":"0a06bffef31dc5f888e335796830f40b7d25388d95911101d91480b1460e1c2e"} Dec 05 10:41:38.995885 master-0 kubenswrapper[10444]: I1205 10:41:38.995686 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:41:38.995885 master-0 kubenswrapper[10444]: I1205 10:41:38.995677 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-2-master-0" event={"ID":"70fbf150-c1f7-4857-8b63-c566282e2526","Type":"ContainerDied","Data":"452fb60a08a0303dfa4a44b4cff2ac5af4105f5fed0d2d7d9442e582b765e031"} Dec 05 10:41:38.995885 master-0 kubenswrapper[10444]: I1205 10:41:38.995742 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="452fb60a08a0303dfa4a44b4cff2ac5af4105f5fed0d2d7d9442e582b765e031" Dec 05 10:41:39.003672 master-0 kubenswrapper[10444]: I1205 10:41:39.003490 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" event={"ID":"0babac16-164a-405a-a86c-30524118f046","Type":"ContainerStarted","Data":"88076357a096b39ba072f983a9ac87d9bec3e707079c1bb405bcd569dd44c1d2"} Dec 05 10:41:39.004176 master-0 kubenswrapper[10444]: I1205 10:41:39.004121 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" Dec 05 10:41:39.014829 master-0 kubenswrapper[10444]: I1205 10:41:39.014782 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kube-system/bootstrap-kube-controller-manager-master-0" Dec 05 10:41:39.015009 master-0 kubenswrapper[10444]: I1205 10:41:39.014893 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-5t4nn" podStartSLOduration=26.014863079 podStartE2EDuration="26.014863079s" podCreationTimestamp="2025-12-05 10:41:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:41:39.011597554 +0000 UTC m=+239.603510141" watchObservedRunningTime="2025-12-05 10:41:39.014863079 +0000 UTC m=+239.606775686" Dec 05 10:41:39.015230 master-0 kubenswrapper[10444]: I1205 10:41:39.015177 10444 scope.go:117] "RemoveContainer" containerID="efee17240223c7e06e121d89cbef5a60b4ccceeebdd17b05f12e6a98b82892f6" Dec 05 10:41:39.015363 master-0 kubenswrapper[10444]: I1205 10:41:39.015315 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" Dec 05 10:41:39.024394 master-0 kubenswrapper[10444]: I1205 10:41:39.024286 10444 generic.go:334] "Generic (PLEG): container finished" podID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerID="e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23" exitCode=0 Dec 05 10:41:39.024563 master-0 kubenswrapper[10444]: I1205 10:41:39.024384 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk29h" event={"ID":"0aa6e138-4b1f-4047-8255-a2b14d044588","Type":"ContainerDied","Data":"e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23"} Dec 05 10:41:39.028116 master-0 kubenswrapper[10444]: I1205 10:41:39.028064 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"a51fa80258077cf0f6fddf343b4ba54f","Type":"ContainerStarted","Data":"b51c377684aafed12957958fdc2a947c34fdd370a6c72a4ef4641c4738cf31b1"} Dec 05 10:41:39.028231 master-0 kubenswrapper[10444]: I1205 10:41:39.028120 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"a51fa80258077cf0f6fddf343b4ba54f","Type":"ContainerStarted","Data":"512588b6c638c1ce7fb1dce16bbe4f74e94e57511137324c00c995274f332c5c"} Dec 05 10:41:39.030773 master-0 kubenswrapper[10444]: I1205 10:41:39.030676 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqhfn" event={"ID":"0a4db531-8af7-4085-a6b2-0de51b527ce6","Type":"ContainerStarted","Data":"ef5e388b6cabad318ee03ac9a678f15288ffd67d2a57ef3139d9a610725fcfd3"} Dec 05 10:41:39.033612 master-0 kubenswrapper[10444]: I1205 10:41:39.033556 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" event={"ID":"445d75af-d072-4fa0-91a7-f3fa579b9ca9","Type":"ContainerStarted","Data":"dfc1c20819ce4f467461d89e6abf68fcdb1202febbeb18490a6b3dcd2aec085f"} Dec 05 10:41:39.035816 master-0 kubenswrapper[10444]: I1205 10:41:39.035769 10444 generic.go:334] "Generic (PLEG): container finished" podID="baa812e4-b011-41ee-82ab-8f571f0b7e0a" containerID="84efa2150952ddba357e673dac2d7ef63c7f6bced1cf84122ed7eed4c4e74f80" exitCode=0 Dec 05 10:41:39.035916 master-0 kubenswrapper[10444]: I1205 10:41:39.035823 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6p8cq" event={"ID":"baa812e4-b011-41ee-82ab-8f571f0b7e0a","Type":"ContainerDied","Data":"84efa2150952ddba357e673dac2d7ef63c7f6bced1cf84122ed7eed4c4e74f80"} Dec 05 10:41:39.072579 master-0 kubenswrapper[10444]: I1205 10:41:39.071598 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" podStartSLOduration=189.681409638 podStartE2EDuration="3m38.071579309s" podCreationTimestamp="2025-12-05 10:38:01 +0000 UTC" firstStartedPulling="2025-12-05 10:41:09.725151636 +0000 UTC m=+210.317064203" lastFinishedPulling="2025-12-05 10:41:38.115321307 +0000 UTC m=+238.707233874" observedRunningTime="2025-12-05 10:41:39.069977258 +0000 UTC m=+239.661889855" watchObservedRunningTime="2025-12-05 10:41:39.071579309 +0000 UTC m=+239.663491886" Dec 05 10:41:39.089736 master-0 kubenswrapper[10444]: I1205 10:41:39.089692 10444 scope.go:117] "RemoveContainer" containerID="80aee350277cce308d84133aa5eefc6be3ed0712557f82d0d22b78dac448f0af" Dec 05 10:41:39.226561 master-0 kubenswrapper[10444]: I1205 10:41:39.226463 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podStartSLOduration=192.732647548 podStartE2EDuration="3m41.226447847s" podCreationTimestamp="2025-12-05 10:37:58 +0000 UTC" firstStartedPulling="2025-12-05 10:41:09.592023498 +0000 UTC m=+210.183936065" lastFinishedPulling="2025-12-05 10:41:38.085823797 +0000 UTC m=+238.677736364" observedRunningTime="2025-12-05 10:41:39.221861838 +0000 UTC m=+239.813774405" watchObservedRunningTime="2025-12-05 10:41:39.226447847 +0000 UTC m=+239.818360414" Dec 05 10:41:39.556397 master-0 kubenswrapper[10444]: I1205 10:41:39.556355 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:39.556523 master-0 kubenswrapper[10444]: I1205 10:41:39.556405 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:41:39.560780 master-0 kubenswrapper[10444]: I1205 10:41:39.560707 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:39.560780 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:39.560780 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:39.560780 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:39.561079 master-0 kubenswrapper[10444]: I1205 10:41:39.560850 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:39.605892 master-0 kubenswrapper[10444]: I1205 10:41:39.605833 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b47694fcc32464ab24d09c23d6efb57" path="/var/lib/kubelet/pods/8b47694fcc32464ab24d09c23d6efb57/volumes" Dec 05 10:41:39.606237 master-0 kubenswrapper[10444]: I1205 10:41:39.606187 10444 mirror_client.go:130] "Deleting a mirror pod" pod="kube-system/bootstrap-kube-controller-manager-master-0" podUID="" Dec 05 10:41:39.621090 master-0 kubenswrapper[10444]: I1205 10:41:39.620999 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["kube-system/bootstrap-kube-controller-manager-master-0"] Dec 05 10:41:39.621090 master-0 kubenswrapper[10444]: I1205 10:41:39.621041 10444 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="kube-system/bootstrap-kube-controller-manager-master-0" mirrorPodUID="bb8fe36f-b78d-43e4-b028-ebd3753d3afe" Dec 05 10:41:39.627525 master-0 kubenswrapper[10444]: I1205 10:41:39.627470 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["kube-system/bootstrap-kube-controller-manager-master-0"] Dec 05 10:41:39.627525 master-0 kubenswrapper[10444]: I1205 10:41:39.627513 10444 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="kube-system/bootstrap-kube-controller-manager-master-0" mirrorPodUID="bb8fe36f-b78d-43e4-b028-ebd3753d3afe" Dec 05 10:41:40.057831 master-0 kubenswrapper[10444]: I1205 10:41:40.057717 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-djhk8" event={"ID":"489b6771-aa16-469f-9883-9a9233d3c379","Type":"ContainerStarted","Data":"d1a680b6b4316e757435a8b7fe1a58e8f83e4acd0375fca743d7a7d511212f8e"} Dec 05 10:41:40.067500 master-0 kubenswrapper[10444]: I1205 10:41:40.067439 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk29h" event={"ID":"0aa6e138-4b1f-4047-8255-a2b14d044588","Type":"ContainerStarted","Data":"2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22"} Dec 05 10:41:40.071324 master-0 kubenswrapper[10444]: I1205 10:41:40.071282 10444 generic.go:334] "Generic (PLEG): container finished" podID="0a4db531-8af7-4085-a6b2-0de51b527ce6" containerID="ef5e388b6cabad318ee03ac9a678f15288ffd67d2a57ef3139d9a610725fcfd3" exitCode=0 Dec 05 10:41:40.071402 master-0 kubenswrapper[10444]: I1205 10:41:40.071338 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqhfn" event={"ID":"0a4db531-8af7-4085-a6b2-0de51b527ce6","Type":"ContainerDied","Data":"ef5e388b6cabad318ee03ac9a678f15288ffd67d2a57ef3139d9a610725fcfd3"} Dec 05 10:41:40.075566 master-0 kubenswrapper[10444]: I1205 10:41:40.075502 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"a51fa80258077cf0f6fddf343b4ba54f","Type":"ContainerStarted","Data":"15c71c8a48ba2bacfb2bd2a3e4d67fbc729bbcb6ae32fc023ae61c538c7d05e9"} Dec 05 10:41:40.075634 master-0 kubenswrapper[10444]: I1205 10:41:40.075581 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"a51fa80258077cf0f6fddf343b4ba54f","Type":"ContainerStarted","Data":"0c2a956260655c961495c5a173ae1d7bbf307d5b8972db5af276ccbded5567cc"} Dec 05 10:41:40.075634 master-0 kubenswrapper[10444]: I1205 10:41:40.075603 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"a51fa80258077cf0f6fddf343b4ba54f","Type":"ContainerStarted","Data":"f19decd176069e2f8d9129c470cba4222beb8bbea5c63430032cfc884efc92be"} Dec 05 10:41:40.078193 master-0 kubenswrapper[10444]: I1205 10:41:40.078148 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6p8cq" event={"ID":"baa812e4-b011-41ee-82ab-8f571f0b7e0a","Type":"ContainerStarted","Data":"ba41c876e2b1255912070890abf792ddb367b49d20d803c88dd8d553b18c1108"} Dec 05 10:41:40.089586 master-0 kubenswrapper[10444]: I1205 10:41:40.089502 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-djhk8" podStartSLOduration=15.226986495 podStartE2EDuration="50.089484889s" podCreationTimestamp="2025-12-05 10:40:50 +0000 UTC" firstStartedPulling="2025-12-05 10:41:04.714818697 +0000 UTC m=+205.306731264" lastFinishedPulling="2025-12-05 10:41:39.577317081 +0000 UTC m=+240.169229658" observedRunningTime="2025-12-05 10:41:40.088582275 +0000 UTC m=+240.680494872" watchObservedRunningTime="2025-12-05 10:41:40.089484889 +0000 UTC m=+240.681397456" Dec 05 10:41:40.119539 master-0 kubenswrapper[10444]: I1205 10:41:40.119410 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podStartSLOduration=21.119373888 podStartE2EDuration="21.119373888s" podCreationTimestamp="2025-12-05 10:41:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:41:40.111500966 +0000 UTC m=+240.703413533" watchObservedRunningTime="2025-12-05 10:41:40.119373888 +0000 UTC m=+240.711286455" Dec 05 10:41:40.133178 master-0 kubenswrapper[10444]: I1205 10:41:40.133062 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wk29h" podStartSLOduration=13.199084316 podStartE2EDuration="48.13303758s" podCreationTimestamp="2025-12-05 10:40:52 +0000 UTC" firstStartedPulling="2025-12-05 10:41:04.610387878 +0000 UTC m=+205.202300445" lastFinishedPulling="2025-12-05 10:41:39.544341142 +0000 UTC m=+240.136253709" observedRunningTime="2025-12-05 10:41:40.130596587 +0000 UTC m=+240.722509154" watchObservedRunningTime="2025-12-05 10:41:40.13303758 +0000 UTC m=+240.724950157" Dec 05 10:41:40.149482 master-0 kubenswrapper[10444]: I1205 10:41:40.149376 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6p8cq" podStartSLOduration=15.176643989 podStartE2EDuration="50.149351561s" podCreationTimestamp="2025-12-05 10:40:50 +0000 UTC" firstStartedPulling="2025-12-05 10:41:04.608031158 +0000 UTC m=+205.199943725" lastFinishedPulling="2025-12-05 10:41:39.58073872 +0000 UTC m=+240.172651297" observedRunningTime="2025-12-05 10:41:40.14816636 +0000 UTC m=+240.740078937" watchObservedRunningTime="2025-12-05 10:41:40.149351561 +0000 UTC m=+240.741264128" Dec 05 10:41:40.558135 master-0 kubenswrapper[10444]: I1205 10:41:40.558081 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:40.558135 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:40.558135 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:40.558135 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:40.558494 master-0 kubenswrapper[10444]: I1205 10:41:40.558148 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:41.085080 master-0 kubenswrapper[10444]: I1205 10:41:41.085026 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pqhfn" event={"ID":"0a4db531-8af7-4085-a6b2-0de51b527ce6","Type":"ContainerStarted","Data":"9c2803e83acc5d1981643a87401f8440fbaf661264a764f36e663e29d4ba23a2"} Dec 05 10:41:41.117373 master-0 kubenswrapper[10444]: I1205 10:41:41.117311 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:41:41.117373 master-0 kubenswrapper[10444]: I1205 10:41:41.117364 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:41:41.190997 master-0 kubenswrapper[10444]: I1205 10:41:41.190904 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pqhfn" podStartSLOduration=13.454460881 podStartE2EDuration="48.190886989s" podCreationTimestamp="2025-12-05 10:40:53 +0000 UTC" firstStartedPulling="2025-12-05 10:41:05.728630981 +0000 UTC m=+206.320543548" lastFinishedPulling="2025-12-05 10:41:40.465057089 +0000 UTC m=+241.056969656" observedRunningTime="2025-12-05 10:41:41.189304858 +0000 UTC m=+241.781217435" watchObservedRunningTime="2025-12-05 10:41:41.190886989 +0000 UTC m=+241.782799556" Dec 05 10:41:41.265184 master-0 kubenswrapper[10444]: I1205 10:41:41.265123 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:41:41.265184 master-0 kubenswrapper[10444]: I1205 10:41:41.265178 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:41:41.559360 master-0 kubenswrapper[10444]: I1205 10:41:41.559297 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:41.559360 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:41.559360 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:41.559360 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:41.559637 master-0 kubenswrapper[10444]: I1205 10:41:41.559381 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:42.072546 master-0 kubenswrapper[10444]: I1205 10:41:42.072357 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:42.072546 master-0 kubenswrapper[10444]: I1205 10:41:42.072467 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:42.072546 master-0 kubenswrapper[10444]: I1205 10:41:42.072501 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:42.072546 master-0 kubenswrapper[10444]: I1205 10:41:42.072526 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:42.077506 master-0 kubenswrapper[10444]: I1205 10:41:42.077443 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:42.080516 master-0 kubenswrapper[10444]: I1205 10:41:42.080409 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:42.163111 master-0 kubenswrapper[10444]: I1205 10:41:42.163053 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-djhk8" podUID="489b6771-aa16-469f-9883-9a9233d3c379" containerName="registry-server" probeResult="failure" output=< Dec 05 10:41:42.163111 master-0 kubenswrapper[10444]: timeout: failed to connect service ":50051" within 1s Dec 05 10:41:42.163111 master-0 kubenswrapper[10444]: > Dec 05 10:41:42.301171 master-0 kubenswrapper[10444]: I1205 10:41:42.301079 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-6p8cq" podUID="baa812e4-b011-41ee-82ab-8f571f0b7e0a" containerName="registry-server" probeResult="failure" output=< Dec 05 10:41:42.301171 master-0 kubenswrapper[10444]: timeout: failed to connect service ":50051" within 1s Dec 05 10:41:42.301171 master-0 kubenswrapper[10444]: > Dec 05 10:41:42.566823 master-0 kubenswrapper[10444]: I1205 10:41:42.566715 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:42.566823 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:42.566823 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:42.566823 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:42.567278 master-0 kubenswrapper[10444]: I1205 10:41:42.566834 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:42.661744 master-0 kubenswrapper[10444]: I1205 10:41:42.661675 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:41:42.662401 master-0 kubenswrapper[10444]: I1205 10:41:42.662371 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:41:42.701928 master-0 kubenswrapper[10444]: I1205 10:41:42.701869 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:41:43.559604 master-0 kubenswrapper[10444]: I1205 10:41:43.559514 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:43.559604 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:43.559604 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:43.559604 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:43.560795 master-0 kubenswrapper[10444]: I1205 10:41:43.559622 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:43.880222 master-0 kubenswrapper[10444]: I1205 10:41:43.880071 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:41:43.880222 master-0 kubenswrapper[10444]: I1205 10:41:43.880145 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:41:44.140626 master-0 kubenswrapper[10444]: I1205 10:41:44.140418 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:41:44.559690 master-0 kubenswrapper[10444]: I1205 10:41:44.559590 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:44.559690 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:44.559690 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:44.559690 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:44.560411 master-0 kubenswrapper[10444]: I1205 10:41:44.559693 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:44.917357 master-0 kubenswrapper[10444]: I1205 10:41:44.917218 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pqhfn" podUID="0a4db531-8af7-4085-a6b2-0de51b527ce6" containerName="registry-server" probeResult="failure" output=< Dec 05 10:41:44.917357 master-0 kubenswrapper[10444]: timeout: failed to connect service ":50051" within 1s Dec 05 10:41:44.917357 master-0 kubenswrapper[10444]: > Dec 05 10:41:45.559606 master-0 kubenswrapper[10444]: I1205 10:41:45.559534 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:45.559606 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:45.559606 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:45.559606 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:45.560635 master-0 kubenswrapper[10444]: I1205 10:41:45.559631 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:46.559567 master-0 kubenswrapper[10444]: I1205 10:41:46.559483 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:46.559567 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:46.559567 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:46.559567 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:46.561092 master-0 kubenswrapper[10444]: I1205 10:41:46.559589 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:47.560054 master-0 kubenswrapper[10444]: I1205 10:41:47.559957 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:47.560054 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:47.560054 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:47.560054 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:47.560054 master-0 kubenswrapper[10444]: I1205 10:41:47.560061 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:48.559918 master-0 kubenswrapper[10444]: I1205 10:41:48.559819 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:48.559918 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:48.559918 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:48.559918 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:48.559918 master-0 kubenswrapper[10444]: I1205 10:41:48.559906 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:49.559106 master-0 kubenswrapper[10444]: I1205 10:41:49.559014 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:49.559106 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:49.559106 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:49.559106 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:49.559106 master-0 kubenswrapper[10444]: I1205 10:41:49.559108 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:50.559553 master-0 kubenswrapper[10444]: I1205 10:41:50.559473 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:50.559553 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:50.559553 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:50.559553 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:50.559553 master-0 kubenswrapper[10444]: I1205 10:41:50.559551 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:51.168658 master-0 kubenswrapper[10444]: I1205 10:41:51.168527 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:41:51.236378 master-0 kubenswrapper[10444]: I1205 10:41:51.236282 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:41:51.316622 master-0 kubenswrapper[10444]: I1205 10:41:51.316562 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:41:51.381784 master-0 kubenswrapper[10444]: I1205 10:41:51.381725 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:41:51.559543 master-0 kubenswrapper[10444]: I1205 10:41:51.559488 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:51.559543 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:51.559543 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:51.559543 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:51.560295 master-0 kubenswrapper[10444]: I1205 10:41:51.560259 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:52.078185 master-0 kubenswrapper[10444]: I1205 10:41:52.078087 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:52.079070 master-0 kubenswrapper[10444]: I1205 10:41:52.079018 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:41:52.558300 master-0 kubenswrapper[10444]: I1205 10:41:52.558220 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:52.558300 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:52.558300 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:52.558300 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:52.558300 master-0 kubenswrapper[10444]: I1205 10:41:52.558285 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:53.557858 master-0 kubenswrapper[10444]: I1205 10:41:53.557807 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:53.557858 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:53.557858 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:53.557858 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:53.557858 master-0 kubenswrapper[10444]: I1205 10:41:53.557853 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:53.921791 master-0 kubenswrapper[10444]: I1205 10:41:53.921665 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:41:53.968066 master-0 kubenswrapper[10444]: I1205 10:41:53.968011 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:41:54.558075 master-0 kubenswrapper[10444]: I1205 10:41:54.557991 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:54.558075 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:54.558075 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:54.558075 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:54.559159 master-0 kubenswrapper[10444]: I1205 10:41:54.558086 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:55.558246 master-0 kubenswrapper[10444]: I1205 10:41:55.558150 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:55.558246 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:55.558246 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:55.558246 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:55.558849 master-0 kubenswrapper[10444]: I1205 10:41:55.558276 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:56.639295 master-0 kubenswrapper[10444]: I1205 10:41:56.639222 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:56.639295 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:56.639295 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:56.639295 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:56.640321 master-0 kubenswrapper[10444]: I1205 10:41:56.639300 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:57.014891 master-0 kubenswrapper[10444]: I1205 10:41:57.014757 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z"] Dec 05 10:41:57.015089 master-0 kubenswrapper[10444]: E1205 10:41:57.015052 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70fbf150-c1f7-4857-8b63-c566282e2526" containerName="installer" Dec 05 10:41:57.015089 master-0 kubenswrapper[10444]: I1205 10:41:57.015072 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="70fbf150-c1f7-4857-8b63-c566282e2526" containerName="installer" Dec 05 10:41:57.015221 master-0 kubenswrapper[10444]: I1205 10:41:57.015192 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="70fbf150-c1f7-4857-8b63-c566282e2526" containerName="installer" Dec 05 10:41:57.015968 master-0 kubenswrapper[10444]: I1205 10:41:57.015934 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.021875 master-0 kubenswrapper[10444]: I1205 10:41:57.021782 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-2jmqj" Dec 05 10:41:57.022012 master-0 kubenswrapper[10444]: I1205 10:41:57.021924 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Dec 05 10:41:57.023313 master-0 kubenswrapper[10444]: I1205 10:41:57.023239 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Dec 05 10:41:57.023401 master-0 kubenswrapper[10444]: I1205 10:41:57.023390 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Dec 05 10:41:57.040370 master-0 kubenswrapper[10444]: I1205 10:41:57.040322 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z"] Dec 05 10:41:57.142727 master-0 kubenswrapper[10444]: I1205 10:41:57.142672 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.142922 master-0 kubenswrapper[10444]: I1205 10:41:57.142755 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.142922 master-0 kubenswrapper[10444]: I1205 10:41:57.142817 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g88bl\" (UniqueName: \"kubernetes.io/projected/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-kube-api-access-g88bl\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.142922 master-0 kubenswrapper[10444]: I1205 10:41:57.142851 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-metrics-client-ca\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.243656 master-0 kubenswrapper[10444]: I1205 10:41:57.243598 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g88bl\" (UniqueName: \"kubernetes.io/projected/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-kube-api-access-g88bl\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.243889 master-0 kubenswrapper[10444]: I1205 10:41:57.243670 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-metrics-client-ca\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.243889 master-0 kubenswrapper[10444]: I1205 10:41:57.243727 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.243889 master-0 kubenswrapper[10444]: I1205 10:41:57.243793 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.244012 master-0 kubenswrapper[10444]: E1205 10:41:57.243936 10444 secret.go:189] Couldn't get secret openshift-monitoring/prometheus-operator-tls: secret "prometheus-operator-tls" not found Dec 05 10:41:57.244012 master-0 kubenswrapper[10444]: E1205 10:41:57.243996 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls podName:b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e nodeName:}" failed. No retries permitted until 2025-12-05 10:41:57.743973035 +0000 UTC m=+258.335885602 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "prometheus-operator-tls" (UniqueName: "kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls") pod "prometheus-operator-6c74d9cb9f-r787z" (UID: "b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e") : secret "prometheus-operator-tls" not found Dec 05 10:41:57.245384 master-0 kubenswrapper[10444]: I1205 10:41:57.245349 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-metrics-client-ca\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.248274 master-0 kubenswrapper[10444]: I1205 10:41:57.248230 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.274763 master-0 kubenswrapper[10444]: I1205 10:41:57.274639 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g88bl\" (UniqueName: \"kubernetes.io/projected/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-kube-api-access-g88bl\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.560158 master-0 kubenswrapper[10444]: I1205 10:41:57.560056 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:57.560158 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:57.560158 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:57.560158 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:57.560158 master-0 kubenswrapper[10444]: I1205 10:41:57.560140 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:57.755979 master-0 kubenswrapper[10444]: I1205 10:41:57.755891 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.766591 master-0 kubenswrapper[10444]: I1205 10:41:57.766538 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:57.973990 master-0 kubenswrapper[10444]: I1205 10:41:57.973783 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:41:58.469507 master-0 kubenswrapper[10444]: I1205 10:41:58.469440 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z"] Dec 05 10:41:58.475633 master-0 kubenswrapper[10444]: W1205 10:41:58.475586 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb02b2ea9_2bd2_4ea8_a96b_16d5f13cf19e.slice/crio-2421d54e72b15809966ab160149e8662f98357fd3da518130cc175e2f9f3791a WatchSource:0}: Error finding container 2421d54e72b15809966ab160149e8662f98357fd3da518130cc175e2f9f3791a: Status 404 returned error can't find the container with id 2421d54e72b15809966ab160149e8662f98357fd3da518130cc175e2f9f3791a Dec 05 10:41:58.558948 master-0 kubenswrapper[10444]: I1205 10:41:58.558869 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:58.558948 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:58.558948 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:58.558948 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:58.558948 master-0 kubenswrapper[10444]: I1205 10:41:58.558951 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:41:59.196921 master-0 kubenswrapper[10444]: I1205 10:41:59.196867 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" event={"ID":"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e","Type":"ContainerStarted","Data":"2421d54e72b15809966ab160149e8662f98357fd3da518130cc175e2f9f3791a"} Dec 05 10:41:59.559447 master-0 kubenswrapper[10444]: I1205 10:41:59.559332 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:41:59.559447 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:41:59.559447 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:41:59.559447 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:41:59.559907 master-0 kubenswrapper[10444]: I1205 10:41:59.559525 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:00.558454 master-0 kubenswrapper[10444]: I1205 10:42:00.558298 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:00.558454 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:00.558454 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:00.558454 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:00.559274 master-0 kubenswrapper[10444]: I1205 10:42:00.558461 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:01.208497 master-0 kubenswrapper[10444]: I1205 10:42:01.208347 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" event={"ID":"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e","Type":"ContainerStarted","Data":"0bba5ba1181a12a76aebca07066b46dd62bea0f4086d1159410d8463d2f55fb6"} Dec 05 10:42:01.208497 master-0 kubenswrapper[10444]: I1205 10:42:01.208405 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" event={"ID":"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e","Type":"ContainerStarted","Data":"489a03ae188362149dff25a7fcccc5401bd87ca3d8a6a6b825a07129aade6ac7"} Dec 05 10:42:01.230541 master-0 kubenswrapper[10444]: I1205 10:42:01.230453 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" podStartSLOduration=3.193127884 podStartE2EDuration="5.230415591s" podCreationTimestamp="2025-12-05 10:41:56 +0000 UTC" firstStartedPulling="2025-12-05 10:41:58.479812177 +0000 UTC m=+259.071724754" lastFinishedPulling="2025-12-05 10:42:00.517099894 +0000 UTC m=+261.109012461" observedRunningTime="2025-12-05 10:42:01.228167753 +0000 UTC m=+261.820080370" watchObservedRunningTime="2025-12-05 10:42:01.230415591 +0000 UTC m=+261.822328158" Dec 05 10:42:01.557970 master-0 kubenswrapper[10444]: I1205 10:42:01.557908 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:01.557970 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:01.557970 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:01.557970 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:01.558289 master-0 kubenswrapper[10444]: I1205 10:42:01.557972 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:02.558380 master-0 kubenswrapper[10444]: I1205 10:42:02.558317 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:02.558380 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:02.558380 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:02.558380 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:02.559092 master-0 kubenswrapper[10444]: I1205 10:42:02.558391 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:03.558825 master-0 kubenswrapper[10444]: I1205 10:42:03.558781 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:03.558825 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:03.558825 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:03.558825 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:03.559554 master-0 kubenswrapper[10444]: I1205 10:42:03.559524 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:03.834190 master-0 kubenswrapper[10444]: I1205 10:42:03.834068 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt"] Dec 05 10:42:03.835170 master-0 kubenswrapper[10444]: I1205 10:42:03.835139 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:03.836717 master-0 kubenswrapper[10444]: I1205 10:42:03.836683 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Dec 05 10:42:03.837184 master-0 kubenswrapper[10444]: I1205 10:42:03.837151 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-x98dp" Dec 05 10:42:03.837345 master-0 kubenswrapper[10444]: I1205 10:42:03.837307 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Dec 05 10:42:03.857231 master-0 kubenswrapper[10444]: I1205 10:42:03.857178 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt"] Dec 05 10:42:03.864052 master-0 kubenswrapper[10444]: I1205 10:42:03.863982 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-bmqsb"] Dec 05 10:42:03.865413 master-0 kubenswrapper[10444]: I1205 10:42:03.865383 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:03.867608 master-0 kubenswrapper[10444]: I1205 10:42:03.867577 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-m2gng" Dec 05 10:42:03.867875 master-0 kubenswrapper[10444]: I1205 10:42:03.867852 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Dec 05 10:42:03.868210 master-0 kubenswrapper[10444]: I1205 10:42:03.868187 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Dec 05 10:42:03.937871 master-0 kubenswrapper[10444]: I1205 10:42:03.937800 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpgqz\" (UniqueName: \"kubernetes.io/projected/d9ff1ae2-8547-42db-9fd0-f4782589ca18-kube-api-access-wpgqz\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:03.937871 master-0 kubenswrapper[10444]: I1205 10:42:03.937857 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-tls\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:03.938302 master-0 kubenswrapper[10444]: I1205 10:42:03.937930 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d9ff1ae2-8547-42db-9fd0-f4782589ca18-metrics-client-ca\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:03.938302 master-0 kubenswrapper[10444]: I1205 10:42:03.937955 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:03.944328 master-0 kubenswrapper[10444]: I1205 10:42:03.943883 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-5857974f64-xj7pj"] Dec 05 10:42:03.946494 master-0 kubenswrapper[10444]: I1205 10:42:03.945234 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:03.955619 master-0 kubenswrapper[10444]: I1205 10:42:03.950704 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-qkdkt" Dec 05 10:42:03.955619 master-0 kubenswrapper[10444]: I1205 10:42:03.950805 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Dec 05 10:42:03.955619 master-0 kubenswrapper[10444]: I1205 10:42:03.951001 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Dec 05 10:42:03.955619 master-0 kubenswrapper[10444]: I1205 10:42:03.951061 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Dec 05 10:42:03.956029 master-0 kubenswrapper[10444]: I1205 10:42:03.955989 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-5857974f64-xj7pj"] Dec 05 10:42:04.038851 master-0 kubenswrapper[10444]: I1205 10:42:04.038763 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.039237 master-0 kubenswrapper[10444]: I1205 10:42:04.038897 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/21a9bd5a-3ade-48ef-8004-e0492daa85ba-metrics-client-ca\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.039237 master-0 kubenswrapper[10444]: I1205 10:42:04.039014 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-tls\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.039237 master-0 kubenswrapper[10444]: I1205 10:42:04.039063 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpgqz\" (UniqueName: \"kubernetes.io/projected/d9ff1ae2-8547-42db-9fd0-f4782589ca18-kube-api-access-wpgqz\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:04.039348 master-0 kubenswrapper[10444]: I1205 10:42:04.039243 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-tls\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:04.039348 master-0 kubenswrapper[10444]: I1205 10:42:04.039278 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-tls\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.039348 master-0 kubenswrapper[10444]: I1205 10:42:04.039299 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-metrics-client-ca\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.039348 master-0 kubenswrapper[10444]: I1205 10:42:04.039322 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-root\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.039348 master-0 kubenswrapper[10444]: I1205 10:42:04.039340 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-textfile\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.039547 master-0 kubenswrapper[10444]: I1205 10:42:04.039392 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.041529 master-0 kubenswrapper[10444]: I1205 10:42:04.039569 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/54f9f9d4-8c31-4636-8010-22ee11b9b323-volume-directive-shadow\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.041529 master-0 kubenswrapper[10444]: I1205 10:42:04.039694 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d9ff1ae2-8547-42db-9fd0-f4782589ca18-metrics-client-ca\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:04.041529 master-0 kubenswrapper[10444]: I1205 10:42:04.039730 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv2tg\" (UniqueName: \"kubernetes.io/projected/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-api-access-sv2tg\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.041529 master-0 kubenswrapper[10444]: I1205 10:42:04.039775 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:04.041529 master-0 kubenswrapper[10444]: I1205 10:42:04.039899 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-sys\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.041529 master-0 kubenswrapper[10444]: I1205 10:42:04.039966 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-wtmp\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.041529 master-0 kubenswrapper[10444]: I1205 10:42:04.040018 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8bpw\" (UniqueName: \"kubernetes.io/projected/21a9bd5a-3ade-48ef-8004-e0492daa85ba-kube-api-access-r8bpw\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.041529 master-0 kubenswrapper[10444]: I1205 10:42:04.040087 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.041529 master-0 kubenswrapper[10444]: I1205 10:42:04.040522 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d9ff1ae2-8547-42db-9fd0-f4782589ca18-metrics-client-ca\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:04.042952 master-0 kubenswrapper[10444]: I1205 10:42:04.042925 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-tls\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:04.045891 master-0 kubenswrapper[10444]: I1205 10:42:04.045852 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:04.054970 master-0 kubenswrapper[10444]: I1205 10:42:04.054926 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpgqz\" (UniqueName: \"kubernetes.io/projected/d9ff1ae2-8547-42db-9fd0-f4782589ca18-kube-api-access-wpgqz\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:04.141450 master-0 kubenswrapper[10444]: I1205 10:42:04.141300 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.141450 master-0 kubenswrapper[10444]: I1205 10:42:04.141357 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/21a9bd5a-3ade-48ef-8004-e0492daa85ba-metrics-client-ca\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.141450 master-0 kubenswrapper[10444]: I1205 10:42:04.141385 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-tls\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.141784 master-0 kubenswrapper[10444]: I1205 10:42:04.141732 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-tls\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.141829 master-0 kubenswrapper[10444]: I1205 10:42:04.141802 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-metrics-client-ca\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.141895 master-0 kubenswrapper[10444]: I1205 10:42:04.141866 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-root\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.141933 master-0 kubenswrapper[10444]: I1205 10:42:04.141901 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-textfile\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.141933 master-0 kubenswrapper[10444]: I1205 10:42:04.141922 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.142076 master-0 kubenswrapper[10444]: I1205 10:42:04.142030 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-root\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.142247 master-0 kubenswrapper[10444]: I1205 10:42:04.142215 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/54f9f9d4-8c31-4636-8010-22ee11b9b323-volume-directive-shadow\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.142336 master-0 kubenswrapper[10444]: I1205 10:42:04.142318 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv2tg\" (UniqueName: \"kubernetes.io/projected/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-api-access-sv2tg\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.142588 master-0 kubenswrapper[10444]: I1205 10:42:04.142513 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-sys\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.142588 master-0 kubenswrapper[10444]: I1205 10:42:04.142572 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-wtmp\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.142708 master-0 kubenswrapper[10444]: I1205 10:42:04.142639 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-sys\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.142779 master-0 kubenswrapper[10444]: I1205 10:42:04.142755 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-wtmp\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.142830 master-0 kubenswrapper[10444]: I1205 10:42:04.142815 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8bpw\" (UniqueName: \"kubernetes.io/projected/21a9bd5a-3ade-48ef-8004-e0492daa85ba-kube-api-access-r8bpw\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.142882 master-0 kubenswrapper[10444]: I1205 10:42:04.142857 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.142882 master-0 kubenswrapper[10444]: I1205 10:42:04.142867 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/54f9f9d4-8c31-4636-8010-22ee11b9b323-volume-directive-shadow\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.142981 master-0 kubenswrapper[10444]: I1205 10:42:04.142956 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/21a9bd5a-3ade-48ef-8004-e0492daa85ba-metrics-client-ca\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.143156 master-0 kubenswrapper[10444]: I1205 10:42:04.143131 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-textfile\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.143791 master-0 kubenswrapper[10444]: I1205 10:42:04.143760 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-metrics-client-ca\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.144161 master-0 kubenswrapper[10444]: I1205 10:42:04.144135 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.145004 master-0 kubenswrapper[10444]: I1205 10:42:04.144963 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-tls\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.145311 master-0 kubenswrapper[10444]: I1205 10:42:04.145268 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.145734 master-0 kubenswrapper[10444]: I1205 10:42:04.145687 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.145996 master-0 kubenswrapper[10444]: I1205 10:42:04.145959 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-tls\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.153800 master-0 kubenswrapper[10444]: I1205 10:42:04.153756 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:42:04.163612 master-0 kubenswrapper[10444]: I1205 10:42:04.163572 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8bpw\" (UniqueName: \"kubernetes.io/projected/21a9bd5a-3ade-48ef-8004-e0492daa85ba-kube-api-access-r8bpw\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.164176 master-0 kubenswrapper[10444]: I1205 10:42:04.164142 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv2tg\" (UniqueName: \"kubernetes.io/projected/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-api-access-sv2tg\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.226004 master-0 kubenswrapper[10444]: I1205 10:42:04.225966 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:42:04.291199 master-0 kubenswrapper[10444]: I1205 10:42:04.291162 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:42:04.559081 master-0 kubenswrapper[10444]: I1205 10:42:04.559029 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:04.559081 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:04.559081 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:04.559081 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:04.559624 master-0 kubenswrapper[10444]: I1205 10:42:04.559102 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:04.573770 master-0 kubenswrapper[10444]: I1205 10:42:04.573718 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt"] Dec 05 10:42:04.579593 master-0 kubenswrapper[10444]: W1205 10:42:04.579405 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9ff1ae2_8547_42db_9fd0_f4782589ca18.slice/crio-4f32a485a101f0d25f8273536aa5f41f281533450b1891e0d6dbaa0dcf7b996e WatchSource:0}: Error finding container 4f32a485a101f0d25f8273536aa5f41f281533450b1891e0d6dbaa0dcf7b996e: Status 404 returned error can't find the container with id 4f32a485a101f0d25f8273536aa5f41f281533450b1891e0d6dbaa0dcf7b996e Dec 05 10:42:04.698564 master-0 kubenswrapper[10444]: I1205 10:42:04.698503 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-5857974f64-xj7pj"] Dec 05 10:42:04.707147 master-0 kubenswrapper[10444]: W1205 10:42:04.707039 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod54f9f9d4_8c31_4636_8010_22ee11b9b323.slice/crio-d996cd9dee5c72648a9d435634d8c36eb7feb019522350be8d7d8f4a1fa79c46 WatchSource:0}: Error finding container d996cd9dee5c72648a9d435634d8c36eb7feb019522350be8d7d8f4a1fa79c46: Status 404 returned error can't find the container with id d996cd9dee5c72648a9d435634d8c36eb7feb019522350be8d7d8f4a1fa79c46 Dec 05 10:42:05.230298 master-0 kubenswrapper[10444]: I1205 10:42:05.230160 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-bmqsb" event={"ID":"21a9bd5a-3ade-48ef-8004-e0492daa85ba","Type":"ContainerStarted","Data":"4cedf7a9bbcb2d72fff4df0c2eb3dda7d05531d0220b30c3f687e88816b7ce46"} Dec 05 10:42:05.230971 master-0 kubenswrapper[10444]: I1205 10:42:05.230940 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" event={"ID":"54f9f9d4-8c31-4636-8010-22ee11b9b323","Type":"ContainerStarted","Data":"d996cd9dee5c72648a9d435634d8c36eb7feb019522350be8d7d8f4a1fa79c46"} Dec 05 10:42:05.232512 master-0 kubenswrapper[10444]: I1205 10:42:05.232475 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" event={"ID":"d9ff1ae2-8547-42db-9fd0-f4782589ca18","Type":"ContainerStarted","Data":"8a8943d3201e6f222b9c4e01fdf8f9c1f9c3c9e03f5e80e103f40497de1b3625"} Dec 05 10:42:05.232605 master-0 kubenswrapper[10444]: I1205 10:42:05.232516 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" event={"ID":"d9ff1ae2-8547-42db-9fd0-f4782589ca18","Type":"ContainerStarted","Data":"47ff4c9d7ea2edefa3b30c3e514d6f77c9b7df1a6019ec6af57750d49a113fa5"} Dec 05 10:42:05.232605 master-0 kubenswrapper[10444]: I1205 10:42:05.232534 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" event={"ID":"d9ff1ae2-8547-42db-9fd0-f4782589ca18","Type":"ContainerStarted","Data":"4f32a485a101f0d25f8273536aa5f41f281533450b1891e0d6dbaa0dcf7b996e"} Dec 05 10:42:05.558995 master-0 kubenswrapper[10444]: I1205 10:42:05.558873 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:05.558995 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:05.558995 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:05.558995 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:05.559955 master-0 kubenswrapper[10444]: I1205 10:42:05.559032 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:06.243409 master-0 kubenswrapper[10444]: I1205 10:42:06.243325 10444 generic.go:334] "Generic (PLEG): container finished" podID="21a9bd5a-3ade-48ef-8004-e0492daa85ba" containerID="8ede0c38534b9efcdf0a87cdc9cd4d9eec594a788625f22f913ffac53544f49c" exitCode=0 Dec 05 10:42:06.243409 master-0 kubenswrapper[10444]: I1205 10:42:06.243381 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-bmqsb" event={"ID":"21a9bd5a-3ade-48ef-8004-e0492daa85ba","Type":"ContainerDied","Data":"8ede0c38534b9efcdf0a87cdc9cd4d9eec594a788625f22f913ffac53544f49c"} Dec 05 10:42:06.560347 master-0 kubenswrapper[10444]: I1205 10:42:06.560293 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:06.560347 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:06.560347 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:06.560347 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:06.561275 master-0 kubenswrapper[10444]: I1205 10:42:06.560359 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:07.252999 master-0 kubenswrapper[10444]: I1205 10:42:07.252895 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-bmqsb" event={"ID":"21a9bd5a-3ade-48ef-8004-e0492daa85ba","Type":"ContainerStarted","Data":"ba83bd507731324eea8f11797b66550f5df9a41c15e95a102ef1a9566ed7b02d"} Dec 05 10:42:07.253298 master-0 kubenswrapper[10444]: I1205 10:42:07.252979 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-bmqsb" event={"ID":"21a9bd5a-3ade-48ef-8004-e0492daa85ba","Type":"ContainerStarted","Data":"97e4bc5b35660071c504d2812d0340a259bd139704144b051d6465f586a4c32a"} Dec 05 10:42:07.255903 master-0 kubenswrapper[10444]: I1205 10:42:07.255851 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" event={"ID":"54f9f9d4-8c31-4636-8010-22ee11b9b323","Type":"ContainerStarted","Data":"d90874dc0d21f1457d200d2a58efbbf2d77c90585fc48644ee4110d58f85b0a5"} Dec 05 10:42:07.256206 master-0 kubenswrapper[10444]: I1205 10:42:07.256163 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" event={"ID":"54f9f9d4-8c31-4636-8010-22ee11b9b323","Type":"ContainerStarted","Data":"d738cc1dab5c2a164268419d63713aa5d054e0fc2b9b92bc5dc087ace62406f1"} Dec 05 10:42:07.256492 master-0 kubenswrapper[10444]: I1205 10:42:07.256387 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" event={"ID":"54f9f9d4-8c31-4636-8010-22ee11b9b323","Type":"ContainerStarted","Data":"0e295f44e5f3aaa191f6cb792eb31743c88963d385a39c89edd1285be186d369"} Dec 05 10:42:07.258728 master-0 kubenswrapper[10444]: I1205 10:42:07.258674 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" event={"ID":"d9ff1ae2-8547-42db-9fd0-f4782589ca18","Type":"ContainerStarted","Data":"8504fce80a535c26f462ccb2883f000f7eca23b5ef66e7acbe697d128730e9c8"} Dec 05 10:42:07.277101 master-0 kubenswrapper[10444]: I1205 10:42:07.277036 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-bmqsb" podStartSLOduration=3.174310719 podStartE2EDuration="4.277015223s" podCreationTimestamp="2025-12-05 10:42:03 +0000 UTC" firstStartedPulling="2025-12-05 10:42:04.255184144 +0000 UTC m=+264.847096711" lastFinishedPulling="2025-12-05 10:42:05.357888628 +0000 UTC m=+265.949801215" observedRunningTime="2025-12-05 10:42:07.275245027 +0000 UTC m=+267.867157614" watchObservedRunningTime="2025-12-05 10:42:07.277015223 +0000 UTC m=+267.868927800" Dec 05 10:42:07.304235 master-0 kubenswrapper[10444]: I1205 10:42:07.304123 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" podStartSLOduration=2.827996453 podStartE2EDuration="4.30409634s" podCreationTimestamp="2025-12-05 10:42:03 +0000 UTC" firstStartedPulling="2025-12-05 10:42:04.861939828 +0000 UTC m=+265.453852395" lastFinishedPulling="2025-12-05 10:42:06.338039715 +0000 UTC m=+266.929952282" observedRunningTime="2025-12-05 10:42:07.29867607 +0000 UTC m=+267.890588677" watchObservedRunningTime="2025-12-05 10:42:07.30409634 +0000 UTC m=+267.896008937" Dec 05 10:42:07.328217 master-0 kubenswrapper[10444]: I1205 10:42:07.328092 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" podStartSLOduration=2.697111322 podStartE2EDuration="4.328073277s" podCreationTimestamp="2025-12-05 10:42:03 +0000 UTC" firstStartedPulling="2025-12-05 10:42:04.711070983 +0000 UTC m=+265.302983540" lastFinishedPulling="2025-12-05 10:42:06.342032928 +0000 UTC m=+266.933945495" observedRunningTime="2025-12-05 10:42:07.324163807 +0000 UTC m=+267.916076384" watchObservedRunningTime="2025-12-05 10:42:07.328073277 +0000 UTC m=+267.919985844" Dec 05 10:42:07.559375 master-0 kubenswrapper[10444]: I1205 10:42:07.559301 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:07.559375 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:07.559375 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:07.559375 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:07.559660 master-0 kubenswrapper[10444]: I1205 10:42:07.559383 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:08.560343 master-0 kubenswrapper[10444]: I1205 10:42:08.560248 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:08.560343 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:08.560343 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:08.560343 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:08.561268 master-0 kubenswrapper[10444]: I1205 10:42:08.560362 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:09.308322 master-0 kubenswrapper[10444]: I1205 10:42:09.308271 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-7c46d76dff-z8d8z"] Dec 05 10:42:09.309013 master-0 kubenswrapper[10444]: I1205 10:42:09.308988 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.312920 master-0 kubenswrapper[10444]: I1205 10:42:09.312874 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Dec 05 10:42:09.312920 master-0 kubenswrapper[10444]: I1205 10:42:09.312911 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Dec 05 10:42:09.313088 master-0 kubenswrapper[10444]: I1205 10:42:09.312968 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-5ll0c5ruaqfm2" Dec 05 10:42:09.313199 master-0 kubenswrapper[10444]: I1205 10:42:09.313175 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Dec 05 10:42:09.313373 master-0 kubenswrapper[10444]: I1205 10:42:09.313350 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Dec 05 10:42:09.314056 master-0 kubenswrapper[10444]: I1205 10:42:09.314026 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-l9vpw" Dec 05 10:42:09.327099 master-0 kubenswrapper[10444]: I1205 10:42:09.327053 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-7c46d76dff-z8d8z"] Dec 05 10:42:09.423186 master-0 kubenswrapper[10444]: I1205 10:42:09.423121 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.423186 master-0 kubenswrapper[10444]: I1205 10:42:09.423203 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.423561 master-0 kubenswrapper[10444]: I1205 10:42:09.423481 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.423640 master-0 kubenswrapper[10444]: I1205 10:42:09.423570 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/3aa9a063-322b-4dc6-a724-05a66026160b-audit-log\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.423640 master-0 kubenswrapper[10444]: I1205 10:42:09.423640 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rlts\" (UniqueName: \"kubernetes.io/projected/3aa9a063-322b-4dc6-a724-05a66026160b-kube-api-access-5rlts\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.423818 master-0 kubenswrapper[10444]: I1205 10:42:09.423776 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.423907 master-0 kubenswrapper[10444]: I1205 10:42:09.423819 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.531452 master-0 kubenswrapper[10444]: I1205 10:42:09.527969 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.531452 master-0 kubenswrapper[10444]: I1205 10:42:09.528053 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/3aa9a063-322b-4dc6-a724-05a66026160b-audit-log\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.531452 master-0 kubenswrapper[10444]: I1205 10:42:09.528103 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rlts\" (UniqueName: \"kubernetes.io/projected/3aa9a063-322b-4dc6-a724-05a66026160b-kube-api-access-5rlts\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.531452 master-0 kubenswrapper[10444]: I1205 10:42:09.528189 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.531452 master-0 kubenswrapper[10444]: I1205 10:42:09.528226 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.531452 master-0 kubenswrapper[10444]: I1205 10:42:09.528305 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.531452 master-0 kubenswrapper[10444]: I1205 10:42:09.528365 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.531452 master-0 kubenswrapper[10444]: I1205 10:42:09.529548 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.531452 master-0 kubenswrapper[10444]: I1205 10:42:09.530756 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.531452 master-0 kubenswrapper[10444]: I1205 10:42:09.531118 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/3aa9a063-322b-4dc6-a724-05a66026160b-audit-log\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.535801 master-0 kubenswrapper[10444]: I1205 10:42:09.535739 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.536243 master-0 kubenswrapper[10444]: I1205 10:42:09.536199 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.537566 master-0 kubenswrapper[10444]: I1205 10:42:09.536696 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.552902 master-0 kubenswrapper[10444]: I1205 10:42:09.552834 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rlts\" (UniqueName: \"kubernetes.io/projected/3aa9a063-322b-4dc6-a724-05a66026160b-kube-api-access-5rlts\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:09.558958 master-0 kubenswrapper[10444]: I1205 10:42:09.558732 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:09.558958 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:09.558958 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:09.558958 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:09.559669 master-0 kubenswrapper[10444]: I1205 10:42:09.559292 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:09.632057 master-0 kubenswrapper[10444]: I1205 10:42:09.632001 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:10.109920 master-0 kubenswrapper[10444]: I1205 10:42:10.109800 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-7c46d76dff-z8d8z"] Dec 05 10:42:10.117039 master-0 kubenswrapper[10444]: W1205 10:42:10.116997 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3aa9a063_322b_4dc6_a724_05a66026160b.slice/crio-9dab351b1866858c81dfc4a1e5ed04a7dcaa8ee50017ec3f0b131de9c613ce86 WatchSource:0}: Error finding container 9dab351b1866858c81dfc4a1e5ed04a7dcaa8ee50017ec3f0b131de9c613ce86: Status 404 returned error can't find the container with id 9dab351b1866858c81dfc4a1e5ed04a7dcaa8ee50017ec3f0b131de9c613ce86 Dec 05 10:42:10.276459 master-0 kubenswrapper[10444]: I1205 10:42:10.276360 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" event={"ID":"3aa9a063-322b-4dc6-a724-05a66026160b","Type":"ContainerStarted","Data":"9dab351b1866858c81dfc4a1e5ed04a7dcaa8ee50017ec3f0b131de9c613ce86"} Dec 05 10:42:10.559159 master-0 kubenswrapper[10444]: I1205 10:42:10.559111 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:10.559159 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:10.559159 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:10.559159 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:10.559530 master-0 kubenswrapper[10444]: I1205 10:42:10.559180 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:11.558531 master-0 kubenswrapper[10444]: I1205 10:42:11.558452 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:11.558531 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:11.558531 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:11.558531 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:11.558531 master-0 kubenswrapper[10444]: I1205 10:42:11.558511 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:12.305915 master-0 kubenswrapper[10444]: I1205 10:42:12.305836 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" event={"ID":"3aa9a063-322b-4dc6-a724-05a66026160b","Type":"ContainerStarted","Data":"20fd7ecb75706506f53b22d6ef8c2f7cefab03481f4f534c87bae4cf23aa6438"} Dec 05 10:42:12.331318 master-0 kubenswrapper[10444]: I1205 10:42:12.331214 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" podStartSLOduration=1.957967452 podStartE2EDuration="3.331177s" podCreationTimestamp="2025-12-05 10:42:09 +0000 UTC" firstStartedPulling="2025-12-05 10:42:10.119476613 +0000 UTC m=+270.711389180" lastFinishedPulling="2025-12-05 10:42:11.492686161 +0000 UTC m=+272.084598728" observedRunningTime="2025-12-05 10:42:12.33076335 +0000 UTC m=+272.922675937" watchObservedRunningTime="2025-12-05 10:42:12.331177 +0000 UTC m=+272.923089567" Dec 05 10:42:12.559206 master-0 kubenswrapper[10444]: I1205 10:42:12.559095 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:12.559206 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:12.559206 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:12.559206 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:12.559206 master-0 kubenswrapper[10444]: I1205 10:42:12.559175 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:13.559349 master-0 kubenswrapper[10444]: I1205 10:42:13.559281 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:13.559349 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:13.559349 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:13.559349 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:13.559349 master-0 kubenswrapper[10444]: I1205 10:42:13.559347 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:13.720833 master-0 kubenswrapper[10444]: I1205 10:42:13.720775 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/installer-3-master-0"] Dec 05 10:42:13.721663 master-0 kubenswrapper[10444]: I1205 10:42:13.721628 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:13.725869 master-0 kubenswrapper[10444]: I1205 10:42:13.725822 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-pcdf6" Dec 05 10:42:13.725986 master-0 kubenswrapper[10444]: I1205 10:42:13.725822 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 05 10:42:13.732728 master-0 kubenswrapper[10444]: I1205 10:42:13.732685 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/installer-3-master-0"] Dec 05 10:42:13.887285 master-0 kubenswrapper[10444]: I1205 10:42:13.887146 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kube-api-access\") pod \"installer-3-master-0\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:13.887285 master-0 kubenswrapper[10444]: I1205 10:42:13.887219 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-var-lock\") pod \"installer-3-master-0\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:13.887929 master-0 kubenswrapper[10444]: I1205 10:42:13.887548 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:13.988900 master-0 kubenswrapper[10444]: I1205 10:42:13.988817 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:13.989099 master-0 kubenswrapper[10444]: I1205 10:42:13.988992 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:13.989099 master-0 kubenswrapper[10444]: I1205 10:42:13.989044 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kube-api-access\") pod \"installer-3-master-0\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:13.989200 master-0 kubenswrapper[10444]: I1205 10:42:13.989148 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-var-lock\") pod \"installer-3-master-0\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:13.989263 master-0 kubenswrapper[10444]: I1205 10:42:13.989232 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-var-lock\") pod \"installer-3-master-0\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:14.006901 master-0 kubenswrapper[10444]: I1205 10:42:14.006834 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kube-api-access\") pod \"installer-3-master-0\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:14.040497 master-0 kubenswrapper[10444]: I1205 10:42:14.040441 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:14.496032 master-0 kubenswrapper[10444]: I1205 10:42:14.495967 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/installer-3-master-0"] Dec 05 10:42:14.558355 master-0 kubenswrapper[10444]: I1205 10:42:14.558270 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:14.558355 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:14.558355 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:14.558355 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:14.558355 master-0 kubenswrapper[10444]: I1205 10:42:14.558344 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:15.326332 master-0 kubenswrapper[10444]: I1205 10:42:15.326278 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-3-master-0" event={"ID":"1b60e9f1-b49e-4804-b56d-857c7bb2a18b","Type":"ContainerStarted","Data":"96138a92afe8bfdf53592444354e7fbbf1f63d3a7a1a9363fd5b53fd0a427c72"} Dec 05 10:42:15.326332 master-0 kubenswrapper[10444]: I1205 10:42:15.326334 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-3-master-0" event={"ID":"1b60e9f1-b49e-4804-b56d-857c7bb2a18b","Type":"ContainerStarted","Data":"690adabd380efbcf30cb261955bc2a3bb79595c5d056733e17f611fb3912f863"} Dec 05 10:42:15.355640 master-0 kubenswrapper[10444]: I1205 10:42:15.355542 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/installer-3-master-0" podStartSLOduration=2.355516093 podStartE2EDuration="2.355516093s" podCreationTimestamp="2025-12-05 10:42:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:42:15.352564277 +0000 UTC m=+275.944476884" watchObservedRunningTime="2025-12-05 10:42:15.355516093 +0000 UTC m=+275.947428670" Dec 05 10:42:15.558729 master-0 kubenswrapper[10444]: I1205 10:42:15.558607 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:15.558729 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:15.558729 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:15.558729 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:15.559197 master-0 kubenswrapper[10444]: I1205 10:42:15.558727 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:16.559282 master-0 kubenswrapper[10444]: I1205 10:42:16.559202 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:16.559282 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:16.559282 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:16.559282 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:16.560045 master-0 kubenswrapper[10444]: I1205 10:42:16.559328 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:17.559377 master-0 kubenswrapper[10444]: I1205 10:42:17.559302 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:17.559377 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:17.559377 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:17.559377 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:17.560038 master-0 kubenswrapper[10444]: I1205 10:42:17.559389 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:18.561081 master-0 kubenswrapper[10444]: I1205 10:42:18.560994 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:18.561081 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:18.561081 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:18.561081 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:18.561081 master-0 kubenswrapper[10444]: I1205 10:42:18.561076 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:19.558984 master-0 kubenswrapper[10444]: I1205 10:42:19.558921 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:19.558984 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:19.558984 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:19.558984 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:19.558984 master-0 kubenswrapper[10444]: I1205 10:42:19.558981 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:20.558037 master-0 kubenswrapper[10444]: I1205 10:42:20.557988 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:20.558037 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:20.558037 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:20.558037 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:20.559157 master-0 kubenswrapper[10444]: I1205 10:42:20.559127 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:21.558643 master-0 kubenswrapper[10444]: I1205 10:42:21.558580 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:21.558643 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:21.558643 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:21.558643 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:21.559314 master-0 kubenswrapper[10444]: I1205 10:42:21.558665 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:22.560578 master-0 kubenswrapper[10444]: I1205 10:42:22.560486 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:22.560578 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:22.560578 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:22.560578 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:22.561944 master-0 kubenswrapper[10444]: I1205 10:42:22.560611 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:23.559307 master-0 kubenswrapper[10444]: I1205 10:42:23.559248 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:23.559307 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:23.559307 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:23.559307 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:23.559616 master-0 kubenswrapper[10444]: I1205 10:42:23.559319 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:24.559363 master-0 kubenswrapper[10444]: I1205 10:42:24.559285 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:24.559363 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:24.559363 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:24.559363 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:24.560414 master-0 kubenswrapper[10444]: I1205 10:42:24.559404 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:25.559795 master-0 kubenswrapper[10444]: I1205 10:42:25.559690 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:25.559795 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:25.559795 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:25.559795 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:25.560819 master-0 kubenswrapper[10444]: I1205 10:42:25.559795 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:26.559750 master-0 kubenswrapper[10444]: I1205 10:42:26.559655 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:26.559750 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:26.559750 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:26.559750 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:26.559750 master-0 kubenswrapper[10444]: I1205 10:42:26.559740 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:27.558959 master-0 kubenswrapper[10444]: I1205 10:42:27.558880 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:27.558959 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:27.558959 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:27.558959 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:27.559277 master-0 kubenswrapper[10444]: I1205 10:42:27.558986 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:28.559532 master-0 kubenswrapper[10444]: I1205 10:42:28.559451 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:28.559532 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:28.559532 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:28.559532 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:28.560291 master-0 kubenswrapper[10444]: I1205 10:42:28.559804 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:29.559345 master-0 kubenswrapper[10444]: I1205 10:42:29.559226 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:29.559345 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:29.559345 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:29.559345 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:29.559345 master-0 kubenswrapper[10444]: I1205 10:42:29.559299 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:29.632472 master-0 kubenswrapper[10444]: I1205 10:42:29.632355 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:29.632472 master-0 kubenswrapper[10444]: I1205 10:42:29.632462 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:30.559044 master-0 kubenswrapper[10444]: I1205 10:42:30.558950 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:30.559044 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:30.559044 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:30.559044 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:30.559492 master-0 kubenswrapper[10444]: I1205 10:42:30.559055 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:30.697575 master-0 kubenswrapper[10444]: I1205 10:42:30.697536 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-knq92"] Dec 05 10:42:30.698992 master-0 kubenswrapper[10444]: I1205 10:42:30.698969 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:42:30.703696 master-0 kubenswrapper[10444]: I1205 10:42:30.703638 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 10:42:30.703910 master-0 kubenswrapper[10444]: I1205 10:42:30.703731 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 10:42:30.703910 master-0 kubenswrapper[10444]: I1205 10:42:30.703739 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-scdp6" Dec 05 10:42:30.705530 master-0 kubenswrapper[10444]: I1205 10:42:30.705492 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 10:42:30.715200 master-0 kubenswrapper[10444]: I1205 10:42:30.715139 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-knq92"] Dec 05 10:42:30.736086 master-0 kubenswrapper[10444]: I1205 10:42:30.736047 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:42:30.736335 master-0 kubenswrapper[10444]: I1205 10:42:30.736316 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgvv6\" (UniqueName: \"kubernetes.io/projected/b7508cd3-7421-4ab9-be1f-318db2853596-kube-api-access-kgvv6\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:42:30.838493 master-0 kubenswrapper[10444]: I1205 10:42:30.838294 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:42:30.838733 master-0 kubenswrapper[10444]: E1205 10:42:30.838534 10444 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: secret "canary-serving-cert" not found Dec 05 10:42:30.838733 master-0 kubenswrapper[10444]: I1205 10:42:30.838595 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgvv6\" (UniqueName: \"kubernetes.io/projected/b7508cd3-7421-4ab9-be1f-318db2853596-kube-api-access-kgvv6\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:42:30.838733 master-0 kubenswrapper[10444]: E1205 10:42:30.838641 10444 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert podName:b7508cd3-7421-4ab9-be1f-318db2853596 nodeName:}" failed. No retries permitted until 2025-12-05 10:42:31.338609934 +0000 UTC m=+291.930522541 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert") pod "ingress-canary-knq92" (UID: "b7508cd3-7421-4ab9-be1f-318db2853596") : secret "canary-serving-cert" not found Dec 05 10:42:30.869460 master-0 kubenswrapper[10444]: I1205 10:42:30.869308 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgvv6\" (UniqueName: \"kubernetes.io/projected/b7508cd3-7421-4ab9-be1f-318db2853596-kube-api-access-kgvv6\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:42:31.346582 master-0 kubenswrapper[10444]: I1205 10:42:31.346498 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:42:31.351275 master-0 kubenswrapper[10444]: I1205 10:42:31.351219 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:42:31.438712 master-0 kubenswrapper[10444]: I1205 10:42:31.438650 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/1.log" Dec 05 10:42:31.439771 master-0 kubenswrapper[10444]: I1205 10:42:31.439724 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/0.log" Dec 05 10:42:31.439851 master-0 kubenswrapper[10444]: I1205 10:42:31.439796 10444 generic.go:334] "Generic (PLEG): container finished" podID="22676fac-b770-4937-9bee-7478bd1babb7" containerID="01ddc03289ad66f2a901b27cad793faf8cd47f3f421b8573890f1eb86f36b80a" exitCode=1 Dec 05 10:42:31.439851 master-0 kubenswrapper[10444]: I1205 10:42:31.439835 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" event={"ID":"22676fac-b770-4937-9bee-7478bd1babb7","Type":"ContainerDied","Data":"01ddc03289ad66f2a901b27cad793faf8cd47f3f421b8573890f1eb86f36b80a"} Dec 05 10:42:31.439946 master-0 kubenswrapper[10444]: I1205 10:42:31.439887 10444 scope.go:117] "RemoveContainer" containerID="7f5edea6bac7d86c19774c19407fa60083727d4b313c43d3272af6d979af6367" Dec 05 10:42:31.440535 master-0 kubenswrapper[10444]: I1205 10:42:31.440490 10444 scope.go:117] "RemoveContainer" containerID="01ddc03289ad66f2a901b27cad793faf8cd47f3f421b8573890f1eb86f36b80a" Dec 05 10:42:31.440831 master-0 kubenswrapper[10444]: E1205 10:42:31.440778 10444 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ingress-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ingress-operator pod=ingress-operator-8649c48786-cgt5x_openshift-ingress-operator(22676fac-b770-4937-9bee-7478bd1babb7)\"" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" podUID="22676fac-b770-4937-9bee-7478bd1babb7" Dec 05 10:42:31.560114 master-0 kubenswrapper[10444]: I1205 10:42:31.559092 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:31.560114 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:31.560114 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:31.560114 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:31.560114 master-0 kubenswrapper[10444]: I1205 10:42:31.559151 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:31.624568 master-0 kubenswrapper[10444]: I1205 10:42:31.624406 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:42:32.056084 master-0 kubenswrapper[10444]: I1205 10:42:32.056010 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-knq92"] Dec 05 10:42:32.446676 master-0 kubenswrapper[10444]: I1205 10:42:32.446503 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-knq92" event={"ID":"b7508cd3-7421-4ab9-be1f-318db2853596","Type":"ContainerStarted","Data":"d5adb6f4beb28db3cf257fa202f340fe4a551b9b25a93d1ab664783a6352b140"} Dec 05 10:42:32.446676 master-0 kubenswrapper[10444]: I1205 10:42:32.446563 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-knq92" event={"ID":"b7508cd3-7421-4ab9-be1f-318db2853596","Type":"ContainerStarted","Data":"aa4c28f3537eda89f3c42c0956c1461303fba100316bf76337e7c6ab1fb2e2e4"} Dec 05 10:42:32.448341 master-0 kubenswrapper[10444]: I1205 10:42:32.448298 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/1.log" Dec 05 10:42:32.466637 master-0 kubenswrapper[10444]: I1205 10:42:32.466571 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-knq92" podStartSLOduration=2.466555741 podStartE2EDuration="2.466555741s" podCreationTimestamp="2025-12-05 10:42:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:42:32.463541823 +0000 UTC m=+293.055454390" watchObservedRunningTime="2025-12-05 10:42:32.466555741 +0000 UTC m=+293.058468308" Dec 05 10:42:32.561158 master-0 kubenswrapper[10444]: I1205 10:42:32.561106 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:32.561158 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:32.561158 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:32.561158 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:32.561685 master-0 kubenswrapper[10444]: I1205 10:42:32.561654 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:33.559146 master-0 kubenswrapper[10444]: I1205 10:42:33.559085 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:33.559146 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:33.559146 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:33.559146 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:33.559759 master-0 kubenswrapper[10444]: I1205 10:42:33.559176 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:34.559029 master-0 kubenswrapper[10444]: I1205 10:42:34.558877 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:34.559029 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:34.559029 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:34.559029 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:34.559029 master-0 kubenswrapper[10444]: I1205 10:42:34.559006 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:35.559692 master-0 kubenswrapper[10444]: I1205 10:42:35.559594 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:35.559692 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:35.559692 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:35.559692 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:35.559692 master-0 kubenswrapper[10444]: I1205 10:42:35.559684 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:36.558679 master-0 kubenswrapper[10444]: I1205 10:42:36.558574 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:36.558679 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:36.558679 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:36.558679 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:36.559138 master-0 kubenswrapper[10444]: I1205 10:42:36.558684 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:37.559144 master-0 kubenswrapper[10444]: I1205 10:42:37.559016 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:37.559144 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:37.559144 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:37.559144 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:37.559144 master-0 kubenswrapper[10444]: I1205 10:42:37.559115 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:38.559884 master-0 kubenswrapper[10444]: I1205 10:42:38.559622 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:38.559884 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:38.559884 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:38.559884 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:38.560536 master-0 kubenswrapper[10444]: I1205 10:42:38.559886 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:39.559203 master-0 kubenswrapper[10444]: I1205 10:42:39.559099 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:39.559203 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:39.559203 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:39.559203 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:39.559203 master-0 kubenswrapper[10444]: I1205 10:42:39.559182 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:39.587965 master-0 kubenswrapper[10444]: I1205 10:42:39.587898 10444 kubelet.go:1505] "Image garbage collection succeeded" Dec 05 10:42:40.252101 master-0 kubenswrapper[10444]: I1205 10:42:40.251964 10444 scope.go:117] "RemoveContainer" containerID="53136dd9eb8d1d61f6ae883e0b488e0fb1c3b2414f554f6f62aa4dc092f5012c" Dec 05 10:42:40.559610 master-0 kubenswrapper[10444]: I1205 10:42:40.559482 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:40.559610 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:40.559610 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:40.559610 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:40.560004 master-0 kubenswrapper[10444]: I1205 10:42:40.559628 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:41.559773 master-0 kubenswrapper[10444]: I1205 10:42:41.559667 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:41.559773 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:41.559773 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:41.559773 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:41.560410 master-0 kubenswrapper[10444]: I1205 10:42:41.559795 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:42.559466 master-0 kubenswrapper[10444]: I1205 10:42:42.559371 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:42.559466 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:42.559466 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:42.559466 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:42.559466 master-0 kubenswrapper[10444]: I1205 10:42:42.559459 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:43.558791 master-0 kubenswrapper[10444]: I1205 10:42:43.558715 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:43.558791 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:43.558791 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:43.558791 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:43.558791 master-0 kubenswrapper[10444]: I1205 10:42:43.558793 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:43.593756 master-0 kubenswrapper[10444]: I1205 10:42:43.593655 10444 scope.go:117] "RemoveContainer" containerID="01ddc03289ad66f2a901b27cad793faf8cd47f3f421b8573890f1eb86f36b80a" Dec 05 10:42:44.538112 master-0 kubenswrapper[10444]: I1205 10:42:44.538052 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/1.log" Dec 05 10:42:44.538841 master-0 kubenswrapper[10444]: I1205 10:42:44.538792 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" event={"ID":"22676fac-b770-4937-9bee-7478bd1babb7","Type":"ContainerStarted","Data":"0eefd6344cec169209984eb638655b84736a398edd25e1b16bc1da6cc4d7c7d9"} Dec 05 10:42:44.560152 master-0 kubenswrapper[10444]: I1205 10:42:44.560097 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:44.560152 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:44.560152 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:44.560152 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:44.560782 master-0 kubenswrapper[10444]: I1205 10:42:44.560157 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:45.560414 master-0 kubenswrapper[10444]: I1205 10:42:45.560353 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:45.560414 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:45.560414 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:45.560414 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:45.561591 master-0 kubenswrapper[10444]: I1205 10:42:45.561538 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:46.558928 master-0 kubenswrapper[10444]: I1205 10:42:46.558874 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:46.558928 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:46.558928 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:46.558928 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:46.559366 master-0 kubenswrapper[10444]: I1205 10:42:46.559335 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:47.558736 master-0 kubenswrapper[10444]: I1205 10:42:47.558668 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:47.558736 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:47.558736 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:47.558736 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:47.558736 master-0 kubenswrapper[10444]: I1205 10:42:47.558738 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:47.999761 master-0 kubenswrapper[10444]: I1205 10:42:47.999592 10444 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-master-0"] Dec 05 10:42:48.000223 master-0 kubenswrapper[10444]: I1205 10:42:48.000130 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager" containerID="cri-o://b51c377684aafed12957958fdc2a947c34fdd370a6c72a4ef4641c4738cf31b1" gracePeriod=30 Dec 05 10:42:48.000223 master-0 kubenswrapper[10444]: I1205 10:42:48.000183 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-cert-syncer" containerID="cri-o://0c2a956260655c961495c5a173ae1d7bbf307d5b8972db5af276ccbded5567cc" gracePeriod=30 Dec 05 10:42:48.000377 master-0 kubenswrapper[10444]: I1205 10:42:48.000237 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="cluster-policy-controller" containerID="cri-o://f19decd176069e2f8d9129c470cba4222beb8bbea5c63430032cfc884efc92be" gracePeriod=30 Dec 05 10:42:48.000377 master-0 kubenswrapper[10444]: I1205 10:42:48.000161 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-recovery-controller" containerID="cri-o://15c71c8a48ba2bacfb2bd2a3e4d67fbc729bbcb6ae32fc023ae61c538c7d05e9" gracePeriod=30 Dec 05 10:42:48.002577 master-0 kubenswrapper[10444]: I1205 10:42:48.002505 10444 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-master-0"] Dec 05 10:42:48.003152 master-0 kubenswrapper[10444]: E1205 10:42:48.003068 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-recovery-controller" Dec 05 10:42:48.003152 master-0 kubenswrapper[10444]: I1205 10:42:48.003140 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-recovery-controller" Dec 05 10:42:48.003295 master-0 kubenswrapper[10444]: E1205 10:42:48.003157 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="cluster-policy-controller" Dec 05 10:42:48.003295 master-0 kubenswrapper[10444]: I1205 10:42:48.003237 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="cluster-policy-controller" Dec 05 10:42:48.003295 master-0 kubenswrapper[10444]: E1205 10:42:48.003260 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager" Dec 05 10:42:48.003295 master-0 kubenswrapper[10444]: I1205 10:42:48.003269 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager" Dec 05 10:42:48.003295 master-0 kubenswrapper[10444]: E1205 10:42:48.003290 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-cert-syncer" Dec 05 10:42:48.003295 master-0 kubenswrapper[10444]: I1205 10:42:48.003299 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-cert-syncer" Dec 05 10:42:48.003675 master-0 kubenswrapper[10444]: I1205 10:42:48.003472 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager" Dec 05 10:42:48.003675 master-0 kubenswrapper[10444]: I1205 10:42:48.003489 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="cluster-policy-controller" Dec 05 10:42:48.003675 master-0 kubenswrapper[10444]: I1205 10:42:48.003506 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-cert-syncer" Dec 05 10:42:48.003675 master-0 kubenswrapper[10444]: I1205 10:42:48.003525 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-recovery-controller" Dec 05 10:42:48.077404 master-0 kubenswrapper[10444]: I1205 10:42:48.077325 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:42:48.077704 master-0 kubenswrapper[10444]: I1205 10:42:48.077569 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:42:48.178546 master-0 kubenswrapper[10444]: I1205 10:42:48.178457 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:42:48.178677 master-0 kubenswrapper[10444]: I1205 10:42:48.178571 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:42:48.178731 master-0 kubenswrapper[10444]: I1205 10:42:48.178692 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:42:48.178879 master-0 kubenswrapper[10444]: I1205 10:42:48.178829 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:42:48.181291 master-0 kubenswrapper[10444]: I1205 10:42:48.181187 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_a51fa80258077cf0f6fddf343b4ba54f/kube-controller-manager-cert-syncer/0.log" Dec 05 10:42:48.182380 master-0 kubenswrapper[10444]: I1205 10:42:48.182326 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:42:48.186909 master-0 kubenswrapper[10444]: I1205 10:42:48.186839 10444 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" oldPodUID="a51fa80258077cf0f6fddf343b4ba54f" podUID="5219435a07a0220d41da97c4fb70abb1" Dec 05 10:42:48.280360 master-0 kubenswrapper[10444]: I1205 10:42:48.280194 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-cert-dir\") pod \"a51fa80258077cf0f6fddf343b4ba54f\" (UID: \"a51fa80258077cf0f6fddf343b4ba54f\") " Dec 05 10:42:48.280360 master-0 kubenswrapper[10444]: I1205 10:42:48.280347 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "a51fa80258077cf0f6fddf343b4ba54f" (UID: "a51fa80258077cf0f6fddf343b4ba54f"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:42:48.280656 master-0 kubenswrapper[10444]: I1205 10:42:48.280359 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-resource-dir\") pod \"a51fa80258077cf0f6fddf343b4ba54f\" (UID: \"a51fa80258077cf0f6fddf343b4ba54f\") " Dec 05 10:42:48.280656 master-0 kubenswrapper[10444]: I1205 10:42:48.280382 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "a51fa80258077cf0f6fddf343b4ba54f" (UID: "a51fa80258077cf0f6fddf343b4ba54f"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:42:48.280656 master-0 kubenswrapper[10444]: I1205 10:42:48.280650 10444 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-cert-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:42:48.280852 master-0 kubenswrapper[10444]: I1205 10:42:48.280665 10444 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a51fa80258077cf0f6fddf343b4ba54f-resource-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:42:48.559319 master-0 kubenswrapper[10444]: I1205 10:42:48.559191 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:48.559319 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:48.559319 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:48.559319 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:48.560280 master-0 kubenswrapper[10444]: I1205 10:42:48.559314 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:48.571336 master-0 kubenswrapper[10444]: I1205 10:42:48.571240 10444 generic.go:334] "Generic (PLEG): container finished" podID="1b60e9f1-b49e-4804-b56d-857c7bb2a18b" containerID="96138a92afe8bfdf53592444354e7fbbf1f63d3a7a1a9363fd5b53fd0a427c72" exitCode=0 Dec 05 10:42:48.571623 master-0 kubenswrapper[10444]: I1205 10:42:48.571399 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-3-master-0" event={"ID":"1b60e9f1-b49e-4804-b56d-857c7bb2a18b","Type":"ContainerDied","Data":"96138a92afe8bfdf53592444354e7fbbf1f63d3a7a1a9363fd5b53fd0a427c72"} Dec 05 10:42:48.575305 master-0 kubenswrapper[10444]: I1205 10:42:48.575232 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_a51fa80258077cf0f6fddf343b4ba54f/kube-controller-manager-cert-syncer/0.log" Dec 05 10:42:48.576464 master-0 kubenswrapper[10444]: I1205 10:42:48.576272 10444 generic.go:334] "Generic (PLEG): container finished" podID="a51fa80258077cf0f6fddf343b4ba54f" containerID="15c71c8a48ba2bacfb2bd2a3e4d67fbc729bbcb6ae32fc023ae61c538c7d05e9" exitCode=0 Dec 05 10:42:48.576464 master-0 kubenswrapper[10444]: I1205 10:42:48.576313 10444 generic.go:334] "Generic (PLEG): container finished" podID="a51fa80258077cf0f6fddf343b4ba54f" containerID="0c2a956260655c961495c5a173ae1d7bbf307d5b8972db5af276ccbded5567cc" exitCode=2 Dec 05 10:42:48.576464 master-0 kubenswrapper[10444]: I1205 10:42:48.576329 10444 generic.go:334] "Generic (PLEG): container finished" podID="a51fa80258077cf0f6fddf343b4ba54f" containerID="f19decd176069e2f8d9129c470cba4222beb8bbea5c63430032cfc884efc92be" exitCode=0 Dec 05 10:42:48.576464 master-0 kubenswrapper[10444]: I1205 10:42:48.576347 10444 generic.go:334] "Generic (PLEG): container finished" podID="a51fa80258077cf0f6fddf343b4ba54f" containerID="b51c377684aafed12957958fdc2a947c34fdd370a6c72a4ef4641c4738cf31b1" exitCode=0 Dec 05 10:42:48.576464 master-0 kubenswrapper[10444]: I1205 10:42:48.576391 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="512588b6c638c1ce7fb1dce16bbe4f74e94e57511137324c00c995274f332c5c" Dec 05 10:42:48.576464 master-0 kubenswrapper[10444]: I1205 10:42:48.576406 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:42:48.597127 master-0 kubenswrapper[10444]: I1205 10:42:48.596993 10444 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" oldPodUID="a51fa80258077cf0f6fddf343b4ba54f" podUID="5219435a07a0220d41da97c4fb70abb1" Dec 05 10:42:48.607903 master-0 kubenswrapper[10444]: I1205 10:42:48.607290 10444 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" oldPodUID="a51fa80258077cf0f6fddf343b4ba54f" podUID="5219435a07a0220d41da97c4fb70abb1" Dec 05 10:42:49.559898 master-0 kubenswrapper[10444]: I1205 10:42:49.559783 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:49.559898 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:49.559898 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:49.559898 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:49.561265 master-0 kubenswrapper[10444]: I1205 10:42:49.559903 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:49.610200 master-0 kubenswrapper[10444]: I1205 10:42:49.610094 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a51fa80258077cf0f6fddf343b4ba54f" path="/var/lib/kubelet/pods/a51fa80258077cf0f6fddf343b4ba54f/volumes" Dec 05 10:42:49.640990 master-0 kubenswrapper[10444]: I1205 10:42:49.640899 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:49.648466 master-0 kubenswrapper[10444]: I1205 10:42:49.647297 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:42:49.942368 master-0 kubenswrapper[10444]: I1205 10:42:49.942038 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:50.005779 master-0 kubenswrapper[10444]: I1205 10:42:50.005700 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kube-api-access\") pod \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " Dec 05 10:42:50.006039 master-0 kubenswrapper[10444]: I1205 10:42:50.005824 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-var-lock\") pod \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " Dec 05 10:42:50.006039 master-0 kubenswrapper[10444]: I1205 10:42:50.005852 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kubelet-dir\") pod \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\" (UID: \"1b60e9f1-b49e-4804-b56d-857c7bb2a18b\") " Dec 05 10:42:50.006039 master-0 kubenswrapper[10444]: I1205 10:42:50.005921 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-var-lock" (OuterVolumeSpecName: "var-lock") pod "1b60e9f1-b49e-4804-b56d-857c7bb2a18b" (UID: "1b60e9f1-b49e-4804-b56d-857c7bb2a18b"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:42:50.006584 master-0 kubenswrapper[10444]: I1205 10:42:50.006039 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "1b60e9f1-b49e-4804-b56d-857c7bb2a18b" (UID: "1b60e9f1-b49e-4804-b56d-857c7bb2a18b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:42:50.006584 master-0 kubenswrapper[10444]: I1205 10:42:50.006392 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:42:50.006584 master-0 kubenswrapper[10444]: I1205 10:42:50.006414 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:42:50.008716 master-0 kubenswrapper[10444]: I1205 10:42:50.008659 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1b60e9f1-b49e-4804-b56d-857c7bb2a18b" (UID: "1b60e9f1-b49e-4804-b56d-857c7bb2a18b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:42:50.108240 master-0 kubenswrapper[10444]: I1205 10:42:50.107766 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1b60e9f1-b49e-4804-b56d-857c7bb2a18b-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:42:50.560126 master-0 kubenswrapper[10444]: I1205 10:42:50.560028 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:50.560126 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:50.560126 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:50.560126 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:50.560796 master-0 kubenswrapper[10444]: I1205 10:42:50.560126 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:50.595899 master-0 kubenswrapper[10444]: I1205 10:42:50.595791 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:42:50.595899 master-0 kubenswrapper[10444]: I1205 10:42:50.595846 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-3-master-0" event={"ID":"1b60e9f1-b49e-4804-b56d-857c7bb2a18b","Type":"ContainerDied","Data":"690adabd380efbcf30cb261955bc2a3bb79595c5d056733e17f611fb3912f863"} Dec 05 10:42:50.596341 master-0 kubenswrapper[10444]: I1205 10:42:50.595921 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="690adabd380efbcf30cb261955bc2a3bb79595c5d056733e17f611fb3912f863" Dec 05 10:42:51.558881 master-0 kubenswrapper[10444]: I1205 10:42:51.558796 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:51.558881 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:51.558881 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:51.558881 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:51.558881 master-0 kubenswrapper[10444]: I1205 10:42:51.558860 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:52.559064 master-0 kubenswrapper[10444]: I1205 10:42:52.558945 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:52.559064 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:52.559064 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:52.559064 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:52.559064 master-0 kubenswrapper[10444]: I1205 10:42:52.559050 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:53.558826 master-0 kubenswrapper[10444]: I1205 10:42:53.558740 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:53.558826 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:53.558826 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:53.558826 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:53.559855 master-0 kubenswrapper[10444]: I1205 10:42:53.558830 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:54.559543 master-0 kubenswrapper[10444]: I1205 10:42:54.559486 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:54.559543 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:54.559543 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:54.559543 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:54.560174 master-0 kubenswrapper[10444]: I1205 10:42:54.559567 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:55.909062 master-0 kubenswrapper[10444]: I1205 10:42:55.909018 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:55.909062 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:55.909062 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:55.909062 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:55.910103 master-0 kubenswrapper[10444]: I1205 10:42:55.910074 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:56.559398 master-0 kubenswrapper[10444]: I1205 10:42:56.559316 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:56.559398 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:56.559398 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:56.559398 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:56.560064 master-0 kubenswrapper[10444]: I1205 10:42:56.559403 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:57.558565 master-0 kubenswrapper[10444]: I1205 10:42:57.558458 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:57.558565 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:57.558565 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:57.558565 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:57.559278 master-0 kubenswrapper[10444]: I1205 10:42:57.558584 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:58.559293 master-0 kubenswrapper[10444]: I1205 10:42:58.559207 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:58.559293 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:58.559293 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:58.559293 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:58.560189 master-0 kubenswrapper[10444]: I1205 10:42:58.559314 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:42:59.558364 master-0 kubenswrapper[10444]: I1205 10:42:59.558305 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:42:59.558364 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:42:59.558364 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:42:59.558364 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:42:59.558654 master-0 kubenswrapper[10444]: I1205 10:42:59.558389 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:00.558716 master-0 kubenswrapper[10444]: I1205 10:43:00.558657 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:00.558716 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:00.558716 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:00.558716 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:00.559628 master-0 kubenswrapper[10444]: I1205 10:43:00.559569 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:00.593996 master-0 kubenswrapper[10444]: I1205 10:43:00.593400 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:43:00.609948 master-0 kubenswrapper[10444]: I1205 10:43:00.609900 10444 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="b918fd1c-9a1b-42b7-85ec-47265ad09ad4" Dec 05 10:43:00.609948 master-0 kubenswrapper[10444]: I1205 10:43:00.609944 10444 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="b918fd1c-9a1b-42b7-85ec-47265ad09ad4" Dec 05 10:43:00.628990 master-0 kubenswrapper[10444]: I1205 10:43:00.628909 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-master-0"] Dec 05 10:43:00.629976 master-0 kubenswrapper[10444]: I1205 10:43:00.629934 10444 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:43:00.633317 master-0 kubenswrapper[10444]: I1205 10:43:00.633275 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-master-0"] Dec 05 10:43:00.645193 master-0 kubenswrapper[10444]: I1205 10:43:00.645151 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:43:00.650741 master-0 kubenswrapper[10444]: I1205 10:43:00.650685 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-master-0"] Dec 05 10:43:00.949829 master-0 kubenswrapper[10444]: I1205 10:43:00.949769 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"a0c4bf77b56c9bf53793c3092e77173ade30a59f4de69720ced91f80c7a365c6"} Dec 05 10:43:00.949829 master-0 kubenswrapper[10444]: I1205 10:43:00.949830 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"5be78d3569389b67ef6d6c6575e7c97a75617917f6e7f31b9b05a692b9e12292"} Dec 05 10:43:01.558701 master-0 kubenswrapper[10444]: I1205 10:43:01.558647 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:01.558701 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:01.558701 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:01.558701 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:01.559614 master-0 kubenswrapper[10444]: I1205 10:43:01.558715 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:01.958020 master-0 kubenswrapper[10444]: I1205 10:43:01.957933 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"f7c86a3fbacaa184c777e4b5c80c29c00ccd9777b6aeed8a46d62880a41cc869"} Dec 05 10:43:01.958020 master-0 kubenswrapper[10444]: I1205 10:43:01.958018 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"47bef9607325414e871a594b06d4250139a33f93aa8b2368ce9838e31d093051"} Dec 05 10:43:01.958261 master-0 kubenswrapper[10444]: I1205 10:43:01.958039 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"00db1aab947f0a6caa3edb56d388eb498e266a62880572bf52a5608eb626a365"} Dec 05 10:43:01.980829 master-0 kubenswrapper[10444]: I1205 10:43:01.980738 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podStartSLOduration=1.9807177839999999 podStartE2EDuration="1.980717784s" podCreationTimestamp="2025-12-05 10:43:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:43:01.979590906 +0000 UTC m=+322.571503473" watchObservedRunningTime="2025-12-05 10:43:01.980717784 +0000 UTC m=+322.572630381" Dec 05 10:43:02.559164 master-0 kubenswrapper[10444]: I1205 10:43:02.559085 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:02.559164 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:02.559164 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:02.559164 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:02.560062 master-0 kubenswrapper[10444]: I1205 10:43:02.559182 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:03.558548 master-0 kubenswrapper[10444]: I1205 10:43:03.558475 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:03.558548 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:03.558548 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:03.558548 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:03.558869 master-0 kubenswrapper[10444]: I1205 10:43:03.558576 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:04.559527 master-0 kubenswrapper[10444]: I1205 10:43:04.559418 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:04.559527 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:04.559527 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:04.559527 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:04.560566 master-0 kubenswrapper[10444]: I1205 10:43:04.559545 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:05.560056 master-0 kubenswrapper[10444]: I1205 10:43:05.559973 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:05.560056 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:05.560056 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:05.560056 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:05.560636 master-0 kubenswrapper[10444]: I1205 10:43:05.560080 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:06.559000 master-0 kubenswrapper[10444]: I1205 10:43:06.558926 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:06.559000 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:06.559000 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:06.559000 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:06.559534 master-0 kubenswrapper[10444]: I1205 10:43:06.559023 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:07.559743 master-0 kubenswrapper[10444]: I1205 10:43:07.559637 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:07.559743 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:07.559743 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:07.559743 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:07.559743 master-0 kubenswrapper[10444]: I1205 10:43:07.559733 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:08.559122 master-0 kubenswrapper[10444]: I1205 10:43:08.559048 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:08.559122 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:08.559122 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:08.559122 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:08.559713 master-0 kubenswrapper[10444]: I1205 10:43:08.559139 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:09.559221 master-0 kubenswrapper[10444]: I1205 10:43:09.559063 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:09.559221 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:09.559221 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:09.559221 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:09.560468 master-0 kubenswrapper[10444]: I1205 10:43:09.559262 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:10.559375 master-0 kubenswrapper[10444]: I1205 10:43:10.559256 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:10.559375 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:10.559375 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:10.559375 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:10.559375 master-0 kubenswrapper[10444]: I1205 10:43:10.559347 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:10.646224 master-0 kubenswrapper[10444]: I1205 10:43:10.646154 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:43:10.646709 master-0 kubenswrapper[10444]: I1205 10:43:10.646665 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:43:10.646960 master-0 kubenswrapper[10444]: I1205 10:43:10.646930 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:43:10.647343 master-0 kubenswrapper[10444]: I1205 10:43:10.647304 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:43:10.652783 master-0 kubenswrapper[10444]: I1205 10:43:10.652706 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:43:10.653782 master-0 kubenswrapper[10444]: I1205 10:43:10.653705 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:43:11.033881 master-0 kubenswrapper[10444]: I1205 10:43:11.033804 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:43:11.034172 master-0 kubenswrapper[10444]: I1205 10:43:11.034128 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:43:11.560375 master-0 kubenswrapper[10444]: I1205 10:43:11.559990 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:11.560375 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:11.560375 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:11.560375 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:11.561031 master-0 kubenswrapper[10444]: I1205 10:43:11.560379 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:12.559223 master-0 kubenswrapper[10444]: I1205 10:43:12.559137 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:12.559223 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:12.559223 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:12.559223 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:12.559631 master-0 kubenswrapper[10444]: I1205 10:43:12.559226 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:13.559060 master-0 kubenswrapper[10444]: I1205 10:43:13.558989 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:13.559060 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:13.559060 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:13.559060 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:13.559636 master-0 kubenswrapper[10444]: I1205 10:43:13.559078 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:14.559975 master-0 kubenswrapper[10444]: I1205 10:43:14.559920 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:14.559975 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:14.559975 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:14.559975 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:14.560560 master-0 kubenswrapper[10444]: I1205 10:43:14.559991 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:15.558719 master-0 kubenswrapper[10444]: I1205 10:43:15.558643 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:15.558719 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:15.558719 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:15.558719 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:15.558719 master-0 kubenswrapper[10444]: I1205 10:43:15.558709 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:16.558753 master-0 kubenswrapper[10444]: I1205 10:43:16.558661 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:16.558753 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:16.558753 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:16.558753 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:16.559374 master-0 kubenswrapper[10444]: I1205 10:43:16.558784 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:17.558255 master-0 kubenswrapper[10444]: I1205 10:43:17.558190 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:17.558255 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:17.558255 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:17.558255 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:17.558654 master-0 kubenswrapper[10444]: I1205 10:43:17.558264 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:18.557667 master-0 kubenswrapper[10444]: I1205 10:43:18.557602 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:18.557667 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:18.557667 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:18.557667 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:18.557667 master-0 kubenswrapper[10444]: I1205 10:43:18.557667 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:19.567740 master-0 kubenswrapper[10444]: I1205 10:43:19.567680 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:19.567740 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:19.567740 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:19.567740 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:19.568396 master-0 kubenswrapper[10444]: I1205 10:43:19.567760 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:20.558687 master-0 kubenswrapper[10444]: I1205 10:43:20.558617 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:20.558687 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:20.558687 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:20.558687 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:20.558687 master-0 kubenswrapper[10444]: I1205 10:43:20.558678 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:21.558688 master-0 kubenswrapper[10444]: I1205 10:43:21.558625 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:21.558688 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:21.558688 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:21.558688 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:21.559228 master-0 kubenswrapper[10444]: I1205 10:43:21.558706 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:22.559205 master-0 kubenswrapper[10444]: I1205 10:43:22.559077 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:22.559205 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:22.559205 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:22.559205 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:22.559205 master-0 kubenswrapper[10444]: I1205 10:43:22.559143 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:23.559807 master-0 kubenswrapper[10444]: I1205 10:43:23.559724 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:23.559807 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:23.559807 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:23.559807 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:23.561194 master-0 kubenswrapper[10444]: I1205 10:43:23.560884 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:24.558263 master-0 kubenswrapper[10444]: I1205 10:43:24.558188 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:24.558263 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:24.558263 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:24.558263 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:24.558783 master-0 kubenswrapper[10444]: I1205 10:43:24.558277 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:25.559545 master-0 kubenswrapper[10444]: I1205 10:43:25.559466 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:25.559545 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:25.559545 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:25.559545 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:25.560199 master-0 kubenswrapper[10444]: I1205 10:43:25.559577 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:26.558661 master-0 kubenswrapper[10444]: I1205 10:43:26.558603 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:26.558661 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:26.558661 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:26.558661 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:26.559095 master-0 kubenswrapper[10444]: I1205 10:43:26.559064 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:27.558712 master-0 kubenswrapper[10444]: I1205 10:43:27.558638 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:27.558712 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:27.558712 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:27.558712 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:27.558712 master-0 kubenswrapper[10444]: I1205 10:43:27.558711 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:28.558939 master-0 kubenswrapper[10444]: I1205 10:43:28.558856 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:28.558939 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:28.558939 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:28.558939 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:28.558939 master-0 kubenswrapper[10444]: I1205 10:43:28.558936 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:29.558967 master-0 kubenswrapper[10444]: I1205 10:43:29.558926 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:29.558967 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:29.558967 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:29.558967 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:29.559639 master-0 kubenswrapper[10444]: I1205 10:43:29.559610 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:30.559202 master-0 kubenswrapper[10444]: I1205 10:43:30.559148 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:30.559202 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:30.559202 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:30.559202 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:30.559818 master-0 kubenswrapper[10444]: I1205 10:43:30.559212 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:31.559808 master-0 kubenswrapper[10444]: I1205 10:43:31.559726 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:31.559808 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:31.559808 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:31.559808 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:31.560441 master-0 kubenswrapper[10444]: I1205 10:43:31.559826 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:32.558286 master-0 kubenswrapper[10444]: I1205 10:43:32.558208 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:32.558286 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:32.558286 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:32.558286 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:32.558649 master-0 kubenswrapper[10444]: I1205 10:43:32.558287 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:33.558052 master-0 kubenswrapper[10444]: I1205 10:43:33.557978 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:33.558052 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:33.558052 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:33.558052 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:33.558052 master-0 kubenswrapper[10444]: I1205 10:43:33.558050 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:34.559061 master-0 kubenswrapper[10444]: I1205 10:43:34.559007 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:34.559061 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:34.559061 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:34.559061 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:34.559599 master-0 kubenswrapper[10444]: I1205 10:43:34.559082 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:35.558317 master-0 kubenswrapper[10444]: I1205 10:43:35.558250 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:35.558317 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:35.558317 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:35.558317 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:35.558890 master-0 kubenswrapper[10444]: I1205 10:43:35.558318 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:36.559540 master-0 kubenswrapper[10444]: I1205 10:43:36.559479 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:36.559540 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:36.559540 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:36.559540 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:36.560412 master-0 kubenswrapper[10444]: I1205 10:43:36.559561 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:37.558904 master-0 kubenswrapper[10444]: I1205 10:43:37.558808 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:37.558904 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:37.558904 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:37.558904 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:37.559183 master-0 kubenswrapper[10444]: I1205 10:43:37.558948 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:38.558702 master-0 kubenswrapper[10444]: I1205 10:43:38.558602 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:43:38.558702 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:43:38.558702 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:43:38.558702 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:43:38.559367 master-0 kubenswrapper[10444]: I1205 10:43:38.558727 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:43:38.559367 master-0 kubenswrapper[10444]: I1205 10:43:38.558812 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:43:38.559977 master-0 kubenswrapper[10444]: I1205 10:43:38.559910 10444 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="router" containerStatusID={"Type":"cri-o","ID":"dfc1c20819ce4f467461d89e6abf68fcdb1202febbeb18490a6b3dcd2aec085f"} pod="openshift-ingress/router-default-5465c8b4db-s4c2f" containerMessage="Container router failed startup probe, will be restarted" Dec 05 10:43:38.560054 master-0 kubenswrapper[10444]: I1205 10:43:38.560013 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" containerID="cri-o://dfc1c20819ce4f467461d89e6abf68fcdb1202febbeb18490a6b3dcd2aec085f" gracePeriod=3600 Dec 05 10:44:25.511449 master-0 kubenswrapper[10444]: I1205 10:44:25.511315 10444 generic.go:334] "Generic (PLEG): container finished" podID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerID="dfc1c20819ce4f467461d89e6abf68fcdb1202febbeb18490a6b3dcd2aec085f" exitCode=0 Dec 05 10:44:25.511449 master-0 kubenswrapper[10444]: I1205 10:44:25.511375 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" event={"ID":"445d75af-d072-4fa0-91a7-f3fa579b9ca9","Type":"ContainerDied","Data":"dfc1c20819ce4f467461d89e6abf68fcdb1202febbeb18490a6b3dcd2aec085f"} Dec 05 10:44:25.511449 master-0 kubenswrapper[10444]: I1205 10:44:25.511403 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" event={"ID":"445d75af-d072-4fa0-91a7-f3fa579b9ca9","Type":"ContainerStarted","Data":"df3d2e1cdc52294902ada6e0e872265f937f9cd41ea5e7643f484cd5a4774ce3"} Dec 05 10:44:25.557140 master-0 kubenswrapper[10444]: I1205 10:44:25.556986 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:44:25.560824 master-0 kubenswrapper[10444]: I1205 10:44:25.560721 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:25.560824 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:25.560824 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:25.560824 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:25.561344 master-0 kubenswrapper[10444]: I1205 10:44:25.560842 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:26.560070 master-0 kubenswrapper[10444]: I1205 10:44:26.559957 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:26.560070 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:26.560070 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:26.560070 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:26.560070 master-0 kubenswrapper[10444]: I1205 10:44:26.560049 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:27.559548 master-0 kubenswrapper[10444]: I1205 10:44:27.559464 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:27.559548 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:27.559548 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:27.559548 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:27.559548 master-0 kubenswrapper[10444]: I1205 10:44:27.559527 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:28.560883 master-0 kubenswrapper[10444]: I1205 10:44:28.560778 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:28.560883 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:28.560883 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:28.560883 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:28.560883 master-0 kubenswrapper[10444]: I1205 10:44:28.560868 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:29.556731 master-0 kubenswrapper[10444]: I1205 10:44:29.556649 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:44:29.559195 master-0 kubenswrapper[10444]: I1205 10:44:29.559139 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:29.559195 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:29.559195 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:29.559195 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:29.559501 master-0 kubenswrapper[10444]: I1205 10:44:29.559199 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:30.560465 master-0 kubenswrapper[10444]: I1205 10:44:30.560358 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:30.560465 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:30.560465 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:30.560465 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:30.561020 master-0 kubenswrapper[10444]: I1205 10:44:30.560495 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:31.559288 master-0 kubenswrapper[10444]: I1205 10:44:31.559214 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:31.559288 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:31.559288 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:31.559288 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:31.559895 master-0 kubenswrapper[10444]: I1205 10:44:31.559293 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:32.617698 master-0 kubenswrapper[10444]: I1205 10:44:32.617621 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:32.617698 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:32.617698 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:32.617698 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:32.618240 master-0 kubenswrapper[10444]: I1205 10:44:32.617731 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:33.559289 master-0 kubenswrapper[10444]: I1205 10:44:33.559215 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:33.559289 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:33.559289 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:33.559289 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:33.559575 master-0 kubenswrapper[10444]: I1205 10:44:33.559309 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:34.561244 master-0 kubenswrapper[10444]: I1205 10:44:34.560977 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:34.561244 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:34.561244 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:34.561244 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:34.562219 master-0 kubenswrapper[10444]: I1205 10:44:34.561366 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:35.560605 master-0 kubenswrapper[10444]: I1205 10:44:35.560518 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:35.560605 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:35.560605 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:35.560605 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:35.560605 master-0 kubenswrapper[10444]: I1205 10:44:35.560604 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:36.559485 master-0 kubenswrapper[10444]: I1205 10:44:36.559400 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:36.559485 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:36.559485 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:36.559485 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:36.560230 master-0 kubenswrapper[10444]: I1205 10:44:36.560191 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:37.559494 master-0 kubenswrapper[10444]: I1205 10:44:37.559343 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:37.559494 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:37.559494 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:37.559494 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:37.561191 master-0 kubenswrapper[10444]: I1205 10:44:37.559489 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:38.560068 master-0 kubenswrapper[10444]: I1205 10:44:38.559950 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:38.560068 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:38.560068 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:38.560068 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:38.560068 master-0 kubenswrapper[10444]: I1205 10:44:38.560057 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:39.558865 master-0 kubenswrapper[10444]: I1205 10:44:39.558777 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:39.558865 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:39.558865 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:39.558865 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:39.559170 master-0 kubenswrapper[10444]: I1205 10:44:39.558888 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:40.559581 master-0 kubenswrapper[10444]: I1205 10:44:40.559499 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:40.559581 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:40.559581 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:40.559581 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:40.560223 master-0 kubenswrapper[10444]: I1205 10:44:40.559602 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:41.559867 master-0 kubenswrapper[10444]: I1205 10:44:41.559771 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:41.559867 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:41.559867 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:41.559867 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:41.559867 master-0 kubenswrapper[10444]: I1205 10:44:41.559862 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:42.560297 master-0 kubenswrapper[10444]: I1205 10:44:42.560211 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:42.560297 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:42.560297 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:42.560297 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:42.561025 master-0 kubenswrapper[10444]: I1205 10:44:42.560316 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:43.559482 master-0 kubenswrapper[10444]: I1205 10:44:43.559338 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:43.559482 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:43.559482 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:43.559482 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:43.559482 master-0 kubenswrapper[10444]: I1205 10:44:43.559443 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:44.558901 master-0 kubenswrapper[10444]: I1205 10:44:44.558808 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:44.558901 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:44.558901 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:44.558901 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:44.559649 master-0 kubenswrapper[10444]: I1205 10:44:44.558915 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:44.714938 master-0 kubenswrapper[10444]: I1205 10:44:44.714807 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/2.log" Dec 05 10:44:44.715595 master-0 kubenswrapper[10444]: I1205 10:44:44.715550 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/1.log" Dec 05 10:44:44.716331 master-0 kubenswrapper[10444]: I1205 10:44:44.716264 10444 generic.go:334] "Generic (PLEG): container finished" podID="22676fac-b770-4937-9bee-7478bd1babb7" containerID="0eefd6344cec169209984eb638655b84736a398edd25e1b16bc1da6cc4d7c7d9" exitCode=1 Dec 05 10:44:44.716456 master-0 kubenswrapper[10444]: I1205 10:44:44.716339 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" event={"ID":"22676fac-b770-4937-9bee-7478bd1babb7","Type":"ContainerDied","Data":"0eefd6344cec169209984eb638655b84736a398edd25e1b16bc1da6cc4d7c7d9"} Dec 05 10:44:44.716456 master-0 kubenswrapper[10444]: I1205 10:44:44.716412 10444 scope.go:117] "RemoveContainer" containerID="01ddc03289ad66f2a901b27cad793faf8cd47f3f421b8573890f1eb86f36b80a" Dec 05 10:44:44.717108 master-0 kubenswrapper[10444]: I1205 10:44:44.717060 10444 scope.go:117] "RemoveContainer" containerID="0eefd6344cec169209984eb638655b84736a398edd25e1b16bc1da6cc4d7c7d9" Dec 05 10:44:44.717322 master-0 kubenswrapper[10444]: E1205 10:44:44.717285 10444 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ingress-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ingress-operator pod=ingress-operator-8649c48786-cgt5x_openshift-ingress-operator(22676fac-b770-4937-9bee-7478bd1babb7)\"" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" podUID="22676fac-b770-4937-9bee-7478bd1babb7" Dec 05 10:44:45.561721 master-0 kubenswrapper[10444]: I1205 10:44:45.561617 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:45.561721 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:45.561721 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:45.561721 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:45.562948 master-0 kubenswrapper[10444]: I1205 10:44:45.561822 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:45.725823 master-0 kubenswrapper[10444]: I1205 10:44:45.725759 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/2.log" Dec 05 10:44:46.558580 master-0 kubenswrapper[10444]: I1205 10:44:46.558500 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:46.558580 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:46.558580 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:46.558580 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:46.558862 master-0 kubenswrapper[10444]: I1205 10:44:46.558623 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:47.560539 master-0 kubenswrapper[10444]: I1205 10:44:47.560341 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:47.560539 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:47.560539 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:47.560539 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:47.560539 master-0 kubenswrapper[10444]: I1205 10:44:47.560490 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:48.559504 master-0 kubenswrapper[10444]: I1205 10:44:48.559408 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:48.559504 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:48.559504 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:48.559504 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:48.559822 master-0 kubenswrapper[10444]: I1205 10:44:48.559517 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:49.560352 master-0 kubenswrapper[10444]: I1205 10:44:49.560264 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:49.560352 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:49.560352 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:49.560352 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:49.560352 master-0 kubenswrapper[10444]: I1205 10:44:49.560367 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:50.559115 master-0 kubenswrapper[10444]: I1205 10:44:50.559031 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:50.559115 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:50.559115 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:50.559115 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:50.559474 master-0 kubenswrapper[10444]: I1205 10:44:50.559148 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:51.559638 master-0 kubenswrapper[10444]: I1205 10:44:51.559576 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:51.559638 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:51.559638 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:51.559638 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:51.560239 master-0 kubenswrapper[10444]: I1205 10:44:51.559657 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:52.559136 master-0 kubenswrapper[10444]: I1205 10:44:52.559064 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:52.559136 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:52.559136 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:52.559136 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:52.559136 master-0 kubenswrapper[10444]: I1205 10:44:52.559128 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:53.560852 master-0 kubenswrapper[10444]: I1205 10:44:53.560734 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:53.560852 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:53.560852 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:53.560852 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:53.561662 master-0 kubenswrapper[10444]: I1205 10:44:53.560843 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:54.559340 master-0 kubenswrapper[10444]: I1205 10:44:54.559256 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:54.559340 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:54.559340 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:54.559340 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:54.559638 master-0 kubenswrapper[10444]: I1205 10:44:54.559358 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:55.562029 master-0 kubenswrapper[10444]: I1205 10:44:55.561876 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:55.562029 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:55.562029 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:55.562029 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:55.562029 master-0 kubenswrapper[10444]: I1205 10:44:55.562013 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:56.559197 master-0 kubenswrapper[10444]: I1205 10:44:56.559111 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:56.559197 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:56.559197 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:56.559197 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:56.559669 master-0 kubenswrapper[10444]: I1205 10:44:56.559214 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:57.559507 master-0 kubenswrapper[10444]: I1205 10:44:57.559392 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:57.559507 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:57.559507 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:57.559507 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:57.560638 master-0 kubenswrapper[10444]: I1205 10:44:57.559542 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:58.559576 master-0 kubenswrapper[10444]: I1205 10:44:58.559497 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:58.559576 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:58.559576 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:58.559576 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:58.560132 master-0 kubenswrapper[10444]: I1205 10:44:58.559625 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:44:58.593734 master-0 kubenswrapper[10444]: I1205 10:44:58.593617 10444 scope.go:117] "RemoveContainer" containerID="0eefd6344cec169209984eb638655b84736a398edd25e1b16bc1da6cc4d7c7d9" Dec 05 10:44:58.594237 master-0 kubenswrapper[10444]: E1205 10:44:58.594155 10444 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ingress-operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ingress-operator pod=ingress-operator-8649c48786-cgt5x_openshift-ingress-operator(22676fac-b770-4937-9bee-7478bd1babb7)\"" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" podUID="22676fac-b770-4937-9bee-7478bd1babb7" Dec 05 10:44:59.560147 master-0 kubenswrapper[10444]: I1205 10:44:59.560033 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:44:59.560147 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:44:59.560147 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:44:59.560147 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:44:59.561173 master-0 kubenswrapper[10444]: I1205 10:44:59.560161 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:00.175942 master-0 kubenswrapper[10444]: I1205 10:45:00.175803 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7"] Dec 05 10:45:00.176135 master-0 kubenswrapper[10444]: E1205 10:45:00.176111 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b60e9f1-b49e-4804-b56d-857c7bb2a18b" containerName="installer" Dec 05 10:45:00.176135 master-0 kubenswrapper[10444]: I1205 10:45:00.176130 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b60e9f1-b49e-4804-b56d-857c7bb2a18b" containerName="installer" Dec 05 10:45:00.176307 master-0 kubenswrapper[10444]: I1205 10:45:00.176285 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b60e9f1-b49e-4804-b56d-857c7bb2a18b" containerName="installer" Dec 05 10:45:00.176886 master-0 kubenswrapper[10444]: I1205 10:45:00.176851 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:00.179449 master-0 kubenswrapper[10444]: I1205 10:45:00.179393 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-fdxl6" Dec 05 10:45:00.179557 master-0 kubenswrapper[10444]: I1205 10:45:00.179505 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 10:45:00.188285 master-0 kubenswrapper[10444]: I1205 10:45:00.188229 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7"] Dec 05 10:45:00.208987 master-0 kubenswrapper[10444]: I1205 10:45:00.208891 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7db4c690-9d13-4c82-9565-f974d554916b-secret-volume\") pod \"collect-profiles-29415525-82cr7\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:00.209224 master-0 kubenswrapper[10444]: I1205 10:45:00.209012 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7db4c690-9d13-4c82-9565-f974d554916b-config-volume\") pod \"collect-profiles-29415525-82cr7\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:00.209224 master-0 kubenswrapper[10444]: I1205 10:45:00.209067 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk8b7\" (UniqueName: \"kubernetes.io/projected/7db4c690-9d13-4c82-9565-f974d554916b-kube-api-access-zk8b7\") pod \"collect-profiles-29415525-82cr7\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:00.310557 master-0 kubenswrapper[10444]: I1205 10:45:00.310390 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7db4c690-9d13-4c82-9565-f974d554916b-secret-volume\") pod \"collect-profiles-29415525-82cr7\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:00.310921 master-0 kubenswrapper[10444]: I1205 10:45:00.310661 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7db4c690-9d13-4c82-9565-f974d554916b-config-volume\") pod \"collect-profiles-29415525-82cr7\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:00.310921 master-0 kubenswrapper[10444]: I1205 10:45:00.310713 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk8b7\" (UniqueName: \"kubernetes.io/projected/7db4c690-9d13-4c82-9565-f974d554916b-kube-api-access-zk8b7\") pod \"collect-profiles-29415525-82cr7\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:00.312160 master-0 kubenswrapper[10444]: I1205 10:45:00.312119 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7db4c690-9d13-4c82-9565-f974d554916b-config-volume\") pod \"collect-profiles-29415525-82cr7\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:00.315655 master-0 kubenswrapper[10444]: I1205 10:45:00.315598 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7db4c690-9d13-4c82-9565-f974d554916b-secret-volume\") pod \"collect-profiles-29415525-82cr7\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:00.327901 master-0 kubenswrapper[10444]: I1205 10:45:00.327836 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk8b7\" (UniqueName: \"kubernetes.io/projected/7db4c690-9d13-4c82-9565-f974d554916b-kube-api-access-zk8b7\") pod \"collect-profiles-29415525-82cr7\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:00.501312 master-0 kubenswrapper[10444]: I1205 10:45:00.501121 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:00.563756 master-0 kubenswrapper[10444]: I1205 10:45:00.563707 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:00.563756 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:00.563756 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:00.563756 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:00.564261 master-0 kubenswrapper[10444]: I1205 10:45:00.563774 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:00.943850 master-0 kubenswrapper[10444]: I1205 10:45:00.943786 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7"] Dec 05 10:45:01.558986 master-0 kubenswrapper[10444]: I1205 10:45:01.558930 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:01.558986 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:01.558986 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:01.558986 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:01.559314 master-0 kubenswrapper[10444]: I1205 10:45:01.559016 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:01.861576 master-0 kubenswrapper[10444]: I1205 10:45:01.861385 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" event={"ID":"7db4c690-9d13-4c82-9565-f974d554916b","Type":"ContainerStarted","Data":"2abcdc2c255198aabeec8ea169e4ae237c27c37467559f6c930cf9b54982525d"} Dec 05 10:45:01.861576 master-0 kubenswrapper[10444]: I1205 10:45:01.861471 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" event={"ID":"7db4c690-9d13-4c82-9565-f974d554916b","Type":"ContainerStarted","Data":"adcd1c5ddd833ada6b59dc1ab4b6b3a5842f80f520ee65ce65a40c8afe637eab"} Dec 05 10:45:01.881107 master-0 kubenswrapper[10444]: I1205 10:45:01.880985 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" podStartSLOduration=1.880962787 podStartE2EDuration="1.880962787s" podCreationTimestamp="2025-12-05 10:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:45:01.879732906 +0000 UTC m=+442.471645483" watchObservedRunningTime="2025-12-05 10:45:01.880962787 +0000 UTC m=+442.472875364" Dec 05 10:45:02.559871 master-0 kubenswrapper[10444]: I1205 10:45:02.559780 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:02.559871 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:02.559871 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:02.559871 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:02.560214 master-0 kubenswrapper[10444]: I1205 10:45:02.559886 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:03.559409 master-0 kubenswrapper[10444]: I1205 10:45:03.559319 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:03.559409 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:03.559409 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:03.559409 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:03.560108 master-0 kubenswrapper[10444]: I1205 10:45:03.559447 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:03.876251 master-0 kubenswrapper[10444]: I1205 10:45:03.876120 10444 generic.go:334] "Generic (PLEG): container finished" podID="7db4c690-9d13-4c82-9565-f974d554916b" containerID="2abcdc2c255198aabeec8ea169e4ae237c27c37467559f6c930cf9b54982525d" exitCode=0 Dec 05 10:45:03.876251 master-0 kubenswrapper[10444]: I1205 10:45:03.876161 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" event={"ID":"7db4c690-9d13-4c82-9565-f974d554916b","Type":"ContainerDied","Data":"2abcdc2c255198aabeec8ea169e4ae237c27c37467559f6c930cf9b54982525d"} Dec 05 10:45:04.558979 master-0 kubenswrapper[10444]: I1205 10:45:04.558889 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:04.558979 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:04.558979 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:04.558979 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:04.558979 master-0 kubenswrapper[10444]: I1205 10:45:04.558964 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:05.231259 master-0 kubenswrapper[10444]: I1205 10:45:05.231192 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:05.391142 master-0 kubenswrapper[10444]: I1205 10:45:05.391050 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7db4c690-9d13-4c82-9565-f974d554916b-config-volume\") pod \"7db4c690-9d13-4c82-9565-f974d554916b\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " Dec 05 10:45:05.391609 master-0 kubenswrapper[10444]: I1205 10:45:05.391291 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zk8b7\" (UniqueName: \"kubernetes.io/projected/7db4c690-9d13-4c82-9565-f974d554916b-kube-api-access-zk8b7\") pod \"7db4c690-9d13-4c82-9565-f974d554916b\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " Dec 05 10:45:05.391609 master-0 kubenswrapper[10444]: I1205 10:45:05.391324 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7db4c690-9d13-4c82-9565-f974d554916b-secret-volume\") pod \"7db4c690-9d13-4c82-9565-f974d554916b\" (UID: \"7db4c690-9d13-4c82-9565-f974d554916b\") " Dec 05 10:45:05.391762 master-0 kubenswrapper[10444]: I1205 10:45:05.391593 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7db4c690-9d13-4c82-9565-f974d554916b-config-volume" (OuterVolumeSpecName: "config-volume") pod "7db4c690-9d13-4c82-9565-f974d554916b" (UID: "7db4c690-9d13-4c82-9565-f974d554916b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:45:05.397545 master-0 kubenswrapper[10444]: I1205 10:45:05.395275 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7db4c690-9d13-4c82-9565-f974d554916b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7db4c690-9d13-4c82-9565-f974d554916b" (UID: "7db4c690-9d13-4c82-9565-f974d554916b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:45:05.397545 master-0 kubenswrapper[10444]: I1205 10:45:05.395273 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7db4c690-9d13-4c82-9565-f974d554916b-kube-api-access-zk8b7" (OuterVolumeSpecName: "kube-api-access-zk8b7") pod "7db4c690-9d13-4c82-9565-f974d554916b" (UID: "7db4c690-9d13-4c82-9565-f974d554916b"). InnerVolumeSpecName "kube-api-access-zk8b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:45:05.493183 master-0 kubenswrapper[10444]: I1205 10:45:05.493001 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zk8b7\" (UniqueName: \"kubernetes.io/projected/7db4c690-9d13-4c82-9565-f974d554916b-kube-api-access-zk8b7\") on node \"master-0\" DevicePath \"\"" Dec 05 10:45:05.493183 master-0 kubenswrapper[10444]: I1205 10:45:05.493088 10444 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7db4c690-9d13-4c82-9565-f974d554916b-secret-volume\") on node \"master-0\" DevicePath \"\"" Dec 05 10:45:05.493183 master-0 kubenswrapper[10444]: I1205 10:45:05.493119 10444 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7db4c690-9d13-4c82-9565-f974d554916b-config-volume\") on node \"master-0\" DevicePath \"\"" Dec 05 10:45:05.560893 master-0 kubenswrapper[10444]: I1205 10:45:05.560772 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:05.560893 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:05.560893 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:05.560893 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:05.561928 master-0 kubenswrapper[10444]: I1205 10:45:05.561604 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:05.902317 master-0 kubenswrapper[10444]: I1205 10:45:05.902246 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" event={"ID":"7db4c690-9d13-4c82-9565-f974d554916b","Type":"ContainerDied","Data":"adcd1c5ddd833ada6b59dc1ab4b6b3a5842f80f520ee65ce65a40c8afe637eab"} Dec 05 10:45:05.902317 master-0 kubenswrapper[10444]: I1205 10:45:05.902307 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adcd1c5ddd833ada6b59dc1ab4b6b3a5842f80f520ee65ce65a40c8afe637eab" Dec 05 10:45:05.902901 master-0 kubenswrapper[10444]: I1205 10:45:05.902849 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:45:06.558874 master-0 kubenswrapper[10444]: I1205 10:45:06.558788 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:06.558874 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:06.558874 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:06.558874 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:06.559325 master-0 kubenswrapper[10444]: I1205 10:45:06.558886 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:07.559457 master-0 kubenswrapper[10444]: I1205 10:45:07.559339 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:07.559457 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:07.559457 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:07.559457 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:07.560505 master-0 kubenswrapper[10444]: I1205 10:45:07.559483 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:08.559369 master-0 kubenswrapper[10444]: I1205 10:45:08.559307 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:08.559369 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:08.559369 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:08.559369 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:08.560627 master-0 kubenswrapper[10444]: I1205 10:45:08.560570 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:09.559866 master-0 kubenswrapper[10444]: I1205 10:45:09.559781 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:09.559866 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:09.559866 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:09.559866 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:09.560614 master-0 kubenswrapper[10444]: I1205 10:45:09.559880 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:10.559222 master-0 kubenswrapper[10444]: I1205 10:45:10.559120 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:10.559222 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:10.559222 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:10.559222 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:10.559222 master-0 kubenswrapper[10444]: I1205 10:45:10.559211 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:11.558222 master-0 kubenswrapper[10444]: I1205 10:45:11.558110 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:11.558222 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:11.558222 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:11.558222 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:11.558946 master-0 kubenswrapper[10444]: I1205 10:45:11.558247 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:12.558535 master-0 kubenswrapper[10444]: I1205 10:45:12.558396 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:12.558535 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:12.558535 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:12.558535 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:12.558535 master-0 kubenswrapper[10444]: I1205 10:45:12.558496 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:13.558520 master-0 kubenswrapper[10444]: I1205 10:45:13.558448 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:13.558520 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:13.558520 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:13.558520 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:13.559136 master-0 kubenswrapper[10444]: I1205 10:45:13.558538 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:13.594676 master-0 kubenswrapper[10444]: I1205 10:45:13.594555 10444 scope.go:117] "RemoveContainer" containerID="0eefd6344cec169209984eb638655b84736a398edd25e1b16bc1da6cc4d7c7d9" Dec 05 10:45:13.966389 master-0 kubenswrapper[10444]: I1205 10:45:13.966251 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/2.log" Dec 05 10:45:13.967017 master-0 kubenswrapper[10444]: I1205 10:45:13.966955 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" event={"ID":"22676fac-b770-4937-9bee-7478bd1babb7","Type":"ContainerStarted","Data":"76b36235b0c1dc04c6934f55253404fa8e8aa924bd0ff655fd03d44b54d22caa"} Dec 05 10:45:14.558876 master-0 kubenswrapper[10444]: I1205 10:45:14.558818 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:14.558876 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:14.558876 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:14.558876 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:14.559399 master-0 kubenswrapper[10444]: I1205 10:45:14.558914 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:15.560312 master-0 kubenswrapper[10444]: I1205 10:45:15.560138 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:15.560312 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:15.560312 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:15.560312 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:15.560312 master-0 kubenswrapper[10444]: I1205 10:45:15.560261 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:16.558959 master-0 kubenswrapper[10444]: I1205 10:45:16.558846 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:16.558959 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:16.558959 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:16.558959 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:16.558959 master-0 kubenswrapper[10444]: I1205 10:45:16.558922 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:17.558723 master-0 kubenswrapper[10444]: I1205 10:45:17.558661 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:17.558723 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:17.558723 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:17.558723 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:17.559267 master-0 kubenswrapper[10444]: I1205 10:45:17.558733 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:18.557943 master-0 kubenswrapper[10444]: I1205 10:45:18.557871 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:18.557943 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:18.557943 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:18.557943 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:18.558219 master-0 kubenswrapper[10444]: I1205 10:45:18.557963 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:19.558461 master-0 kubenswrapper[10444]: I1205 10:45:19.558376 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:19.558461 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:19.558461 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:19.558461 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:19.559039 master-0 kubenswrapper[10444]: I1205 10:45:19.558503 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:20.559640 master-0 kubenswrapper[10444]: I1205 10:45:20.559551 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:20.559640 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:20.559640 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:20.559640 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:20.559640 master-0 kubenswrapper[10444]: I1205 10:45:20.559634 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:21.557835 master-0 kubenswrapper[10444]: I1205 10:45:21.557779 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:21.557835 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:21.557835 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:21.557835 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:21.558163 master-0 kubenswrapper[10444]: I1205 10:45:21.557840 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:21.903740 master-0 kubenswrapper[10444]: I1205 10:45:21.903574 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/installer-4-master-0"] Dec 05 10:45:21.904799 master-0 kubenswrapper[10444]: E1205 10:45:21.903840 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db4c690-9d13-4c82-9565-f974d554916b" containerName="collect-profiles" Dec 05 10:45:21.904799 master-0 kubenswrapper[10444]: I1205 10:45:21.903870 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db4c690-9d13-4c82-9565-f974d554916b" containerName="collect-profiles" Dec 05 10:45:21.904799 master-0 kubenswrapper[10444]: I1205 10:45:21.903974 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db4c690-9d13-4c82-9565-f974d554916b" containerName="collect-profiles" Dec 05 10:45:21.904799 master-0 kubenswrapper[10444]: I1205 10:45:21.904370 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:21.909856 master-0 kubenswrapper[10444]: I1205 10:45:21.909768 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler"/"installer-sa-dockercfg-v25ks" Dec 05 10:45:21.919078 master-0 kubenswrapper[10444]: I1205 10:45:21.917104 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler"/"kube-root-ca.crt" Dec 05 10:45:21.919078 master-0 kubenswrapper[10444]: I1205 10:45:21.918039 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b3c5922c-65b0-4c75-992c-f9c88f81762d-kube-api-access\") pod \"installer-4-master-0\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:21.919078 master-0 kubenswrapper[10444]: I1205 10:45:21.918138 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-kubelet-dir\") pod \"installer-4-master-0\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:21.919078 master-0 kubenswrapper[10444]: I1205 10:45:21.918174 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-var-lock\") pod \"installer-4-master-0\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:21.922312 master-0 kubenswrapper[10444]: I1205 10:45:21.922221 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/installer-4-master-0"] Dec 05 10:45:22.020954 master-0 kubenswrapper[10444]: I1205 10:45:22.020833 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b3c5922c-65b0-4c75-992c-f9c88f81762d-kube-api-access\") pod \"installer-4-master-0\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:22.021296 master-0 kubenswrapper[10444]: I1205 10:45:22.021162 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-kubelet-dir\") pod \"installer-4-master-0\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:22.021296 master-0 kubenswrapper[10444]: I1205 10:45:22.021231 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-var-lock\") pod \"installer-4-master-0\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:22.021479 master-0 kubenswrapper[10444]: I1205 10:45:22.021370 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-var-lock\") pod \"installer-4-master-0\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:22.021479 master-0 kubenswrapper[10444]: I1205 10:45:22.021408 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-kubelet-dir\") pod \"installer-4-master-0\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:22.039072 master-0 kubenswrapper[10444]: I1205 10:45:22.039006 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b3c5922c-65b0-4c75-992c-f9c88f81762d-kube-api-access\") pod \"installer-4-master-0\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:22.237673 master-0 kubenswrapper[10444]: I1205 10:45:22.237449 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:22.559798 master-0 kubenswrapper[10444]: I1205 10:45:22.559710 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:22.559798 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:22.559798 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:22.559798 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:22.560112 master-0 kubenswrapper[10444]: I1205 10:45:22.559811 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:22.635362 master-0 kubenswrapper[10444]: I1205 10:45:22.635263 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/installer-4-master-0"] Dec 05 10:45:23.022219 master-0 kubenswrapper[10444]: I1205 10:45:23.022171 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-4-master-0" event={"ID":"b3c5922c-65b0-4c75-992c-f9c88f81762d","Type":"ContainerStarted","Data":"f02bbe84a0ab6b5c3f31276301465d91b619a21b309cbdf880fc6c1df1dcf605"} Dec 05 10:45:23.022219 master-0 kubenswrapper[10444]: I1205 10:45:23.022230 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-4-master-0" event={"ID":"b3c5922c-65b0-4c75-992c-f9c88f81762d","Type":"ContainerStarted","Data":"4293a095feecc2a652807fec9aa96d59e590e18262f101aef59da2ccb3b4fbf6"} Dec 05 10:45:23.047592 master-0 kubenswrapper[10444]: I1205 10:45:23.047517 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/installer-4-master-0" podStartSLOduration=2.047501788 podStartE2EDuration="2.047501788s" podCreationTimestamp="2025-12-05 10:45:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:45:23.043776265 +0000 UTC m=+463.635688852" watchObservedRunningTime="2025-12-05 10:45:23.047501788 +0000 UTC m=+463.639414355" Dec 05 10:45:23.559381 master-0 kubenswrapper[10444]: I1205 10:45:23.559313 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:23.559381 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:23.559381 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:23.559381 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:23.559766 master-0 kubenswrapper[10444]: I1205 10:45:23.559467 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:24.559166 master-0 kubenswrapper[10444]: I1205 10:45:24.559098 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:24.559166 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:24.559166 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:24.559166 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:24.559812 master-0 kubenswrapper[10444]: I1205 10:45:24.559175 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:25.558098 master-0 kubenswrapper[10444]: I1205 10:45:25.558010 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:25.558098 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:25.558098 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:25.558098 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:25.558449 master-0 kubenswrapper[10444]: I1205 10:45:25.558114 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:26.558349 master-0 kubenswrapper[10444]: I1205 10:45:26.558290 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:26.558349 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:26.558349 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:26.558349 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:26.558349 master-0 kubenswrapper[10444]: I1205 10:45:26.558351 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:27.559482 master-0 kubenswrapper[10444]: I1205 10:45:27.559346 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:27.559482 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:27.559482 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:27.559482 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:27.560818 master-0 kubenswrapper[10444]: I1205 10:45:27.559503 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:28.559010 master-0 kubenswrapper[10444]: I1205 10:45:28.558947 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:28.559010 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:28.559010 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:28.559010 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:28.559271 master-0 kubenswrapper[10444]: I1205 10:45:28.559028 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:29.558528 master-0 kubenswrapper[10444]: I1205 10:45:29.558457 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:29.558528 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:29.558528 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:29.558528 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:29.558528 master-0 kubenswrapper[10444]: I1205 10:45:29.558521 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:30.559350 master-0 kubenswrapper[10444]: I1205 10:45:30.559277 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:30.559350 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:30.559350 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:30.559350 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:30.559939 master-0 kubenswrapper[10444]: I1205 10:45:30.559369 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:31.558525 master-0 kubenswrapper[10444]: I1205 10:45:31.558458 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:31.558525 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:31.558525 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:31.558525 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:31.558811 master-0 kubenswrapper[10444]: I1205 10:45:31.558529 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:32.558353 master-0 kubenswrapper[10444]: I1205 10:45:32.558288 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:32.558353 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:32.558353 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:32.558353 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:32.559031 master-0 kubenswrapper[10444]: I1205 10:45:32.559005 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:33.558720 master-0 kubenswrapper[10444]: I1205 10:45:33.558651 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:33.558720 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:33.558720 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:33.558720 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:33.559500 master-0 kubenswrapper[10444]: I1205 10:45:33.558740 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:34.558806 master-0 kubenswrapper[10444]: I1205 10:45:34.558729 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:34.558806 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:34.558806 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:34.558806 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:34.559397 master-0 kubenswrapper[10444]: I1205 10:45:34.558808 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:35.560316 master-0 kubenswrapper[10444]: I1205 10:45:35.560243 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:35.560316 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:35.560316 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:35.560316 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:35.561267 master-0 kubenswrapper[10444]: I1205 10:45:35.560333 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:36.560341 master-0 kubenswrapper[10444]: I1205 10:45:36.560224 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:36.560341 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:36.560341 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:36.560341 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:36.560341 master-0 kubenswrapper[10444]: I1205 10:45:36.560321 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:37.559630 master-0 kubenswrapper[10444]: I1205 10:45:37.559527 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:37.559630 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:37.559630 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:37.559630 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:37.560224 master-0 kubenswrapper[10444]: I1205 10:45:37.559644 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:38.559006 master-0 kubenswrapper[10444]: I1205 10:45:38.558928 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:38.559006 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:38.559006 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:38.559006 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:38.559006 master-0 kubenswrapper[10444]: I1205 10:45:38.559004 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:39.558316 master-0 kubenswrapper[10444]: I1205 10:45:39.558269 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:39.558316 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:39.558316 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:39.558316 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:39.558720 master-0 kubenswrapper[10444]: I1205 10:45:39.558323 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:40.561150 master-0 kubenswrapper[10444]: I1205 10:45:40.561092 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:40.561150 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:40.561150 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:40.561150 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:40.561942 master-0 kubenswrapper[10444]: I1205 10:45:40.561167 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:41.560121 master-0 kubenswrapper[10444]: I1205 10:45:41.560030 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:41.560121 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:41.560121 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:41.560121 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:41.560642 master-0 kubenswrapper[10444]: I1205 10:45:41.560125 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:42.559210 master-0 kubenswrapper[10444]: I1205 10:45:42.559103 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:42.559210 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:42.559210 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:42.559210 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:42.559210 master-0 kubenswrapper[10444]: I1205 10:45:42.559191 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:43.559299 master-0 kubenswrapper[10444]: I1205 10:45:43.559257 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:43.559299 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:43.559299 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:43.559299 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:43.559933 master-0 kubenswrapper[10444]: I1205 10:45:43.559904 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:44.560288 master-0 kubenswrapper[10444]: I1205 10:45:44.560205 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:44.560288 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:44.560288 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:44.560288 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:44.561340 master-0 kubenswrapper[10444]: I1205 10:45:44.560311 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:45.559832 master-0 kubenswrapper[10444]: I1205 10:45:45.559717 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:45.559832 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:45.559832 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:45.559832 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:45.559832 master-0 kubenswrapper[10444]: I1205 10:45:45.559799 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:46.559935 master-0 kubenswrapper[10444]: I1205 10:45:46.559826 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:46.559935 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:46.559935 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:46.559935 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:46.560231 master-0 kubenswrapper[10444]: I1205 10:45:46.559977 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:47.558679 master-0 kubenswrapper[10444]: I1205 10:45:47.558586 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:47.558679 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:47.558679 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:47.558679 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:47.558679 master-0 kubenswrapper[10444]: I1205 10:45:47.558665 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:48.558104 master-0 kubenswrapper[10444]: I1205 10:45:48.558027 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:48.558104 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:48.558104 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:48.558104 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:48.558443 master-0 kubenswrapper[10444]: I1205 10:45:48.558126 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:49.338579 master-0 kubenswrapper[10444]: I1205 10:45:49.338504 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-m42rr"] Dec 05 10:45:49.339288 master-0 kubenswrapper[10444]: I1205 10:45:49.339255 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.341568 master-0 kubenswrapper[10444]: I1205 10:45:49.341515 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-d6ph2" Dec 05 10:45:49.341966 master-0 kubenswrapper[10444]: I1205 10:45:49.341938 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Dec 05 10:45:49.492297 master-0 kubenswrapper[10444]: I1205 10:45:49.492207 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.492297 master-0 kubenswrapper[10444]: I1205 10:45:49.492301 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.492744 master-0 kubenswrapper[10444]: I1205 10:45:49.492332 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-ready\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.492744 master-0 kubenswrapper[10444]: I1205 10:45:49.492613 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5ld6\" (UniqueName: \"kubernetes.io/projected/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-kube-api-access-s5ld6\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.559367 master-0 kubenswrapper[10444]: I1205 10:45:49.559294 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:49.559367 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:49.559367 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:49.559367 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:49.559706 master-0 kubenswrapper[10444]: I1205 10:45:49.559384 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:49.593868 master-0 kubenswrapper[10444]: I1205 10:45:49.593724 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.593868 master-0 kubenswrapper[10444]: I1205 10:45:49.593805 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.594105 master-0 kubenswrapper[10444]: I1205 10:45:49.594035 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-ready\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.594214 master-0 kubenswrapper[10444]: I1205 10:45:49.594173 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.594476 master-0 kubenswrapper[10444]: I1205 10:45:49.594378 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5ld6\" (UniqueName: \"kubernetes.io/projected/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-kube-api-access-s5ld6\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.594997 master-0 kubenswrapper[10444]: I1205 10:45:49.594924 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.595060 master-0 kubenswrapper[10444]: I1205 10:45:49.594930 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-ready\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.615415 master-0 kubenswrapper[10444]: I1205 10:45:49.615332 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5ld6\" (UniqueName: \"kubernetes.io/projected/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-kube-api-access-s5ld6\") pod \"cni-sysctl-allowlist-ds-m42rr\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.661308 master-0 kubenswrapper[10444]: I1205 10:45:49.661208 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:49.681857 master-0 kubenswrapper[10444]: W1205 10:45:49.681630 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3034770_d607_4aa2_9b4b_cf1e4b7acf4c.slice/crio-bfef724a46fd2289f7233d41b2c5228c9aaaf92dcf1427768b977e1adf40175a WatchSource:0}: Error finding container bfef724a46fd2289f7233d41b2c5228c9aaaf92dcf1427768b977e1adf40175a: Status 404 returned error can't find the container with id bfef724a46fd2289f7233d41b2c5228c9aaaf92dcf1427768b977e1adf40175a Dec 05 10:45:50.231256 master-0 kubenswrapper[10444]: I1205 10:45:50.231123 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" event={"ID":"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c","Type":"ContainerStarted","Data":"97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30"} Dec 05 10:45:50.231256 master-0 kubenswrapper[10444]: I1205 10:45:50.231213 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" event={"ID":"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c","Type":"ContainerStarted","Data":"bfef724a46fd2289f7233d41b2c5228c9aaaf92dcf1427768b977e1adf40175a"} Dec 05 10:45:50.558411 master-0 kubenswrapper[10444]: I1205 10:45:50.558299 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:50.558411 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:50.558411 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:50.558411 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:50.559334 master-0 kubenswrapper[10444]: I1205 10:45:50.558409 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:51.560167 master-0 kubenswrapper[10444]: I1205 10:45:51.560042 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:51.560167 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:51.560167 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:51.560167 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:51.561298 master-0 kubenswrapper[10444]: I1205 10:45:51.560185 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:52.249134 master-0 kubenswrapper[10444]: I1205 10:45:52.249032 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:52.283742 master-0 kubenswrapper[10444]: I1205 10:45:52.283647 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:45:52.560304 master-0 kubenswrapper[10444]: I1205 10:45:52.560174 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:52.560304 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:52.560304 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:52.560304 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:52.560304 master-0 kubenswrapper[10444]: I1205 10:45:52.560296 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:53.558559 master-0 kubenswrapper[10444]: I1205 10:45:53.558481 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:53.558559 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:53.558559 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:53.558559 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:53.558559 master-0 kubenswrapper[10444]: I1205 10:45:53.558556 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:54.560145 master-0 kubenswrapper[10444]: I1205 10:45:54.560050 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:54.560145 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:54.560145 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:54.560145 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:54.560145 master-0 kubenswrapper[10444]: I1205 10:45:54.560146 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:54.721537 master-0 kubenswrapper[10444]: I1205 10:45:54.721334 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" podStartSLOduration=5.721302887 podStartE2EDuration="5.721302887s" podCreationTimestamp="2025-12-05 10:45:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:45:54.686149738 +0000 UTC m=+495.278062345" watchObservedRunningTime="2025-12-05 10:45:54.721302887 +0000 UTC m=+495.313215494" Dec 05 10:45:54.721811 master-0 kubenswrapper[10444]: I1205 10:45:54.721613 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-1-retry-1-master-0"] Dec 05 10:45:54.722917 master-0 kubenswrapper[10444]: I1205 10:45:54.722863 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:45:54.725452 master-0 kubenswrapper[10444]: I1205 10:45:54.725388 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-grfdw" Dec 05 10:45:54.725929 master-0 kubenswrapper[10444]: I1205 10:45:54.725906 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 10:45:54.736408 master-0 kubenswrapper[10444]: I1205 10:45:54.736355 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-1-retry-1-master-0"] Dec 05 10:45:54.873836 master-0 kubenswrapper[10444]: I1205 10:45:54.873700 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-var-lock\") pod \"installer-1-retry-1-master-0\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:45:54.873836 master-0 kubenswrapper[10444]: I1205 10:45:54.873755 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d1a50dc7-2abb-4183-a391-6f75ab05110c-kube-api-access\") pod \"installer-1-retry-1-master-0\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:45:54.873836 master-0 kubenswrapper[10444]: I1205 10:45:54.873784 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-kubelet-dir\") pod \"installer-1-retry-1-master-0\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:45:54.975406 master-0 kubenswrapper[10444]: I1205 10:45:54.975321 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-var-lock\") pod \"installer-1-retry-1-master-0\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:45:54.975406 master-0 kubenswrapper[10444]: I1205 10:45:54.975384 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d1a50dc7-2abb-4183-a391-6f75ab05110c-kube-api-access\") pod \"installer-1-retry-1-master-0\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:45:54.975406 master-0 kubenswrapper[10444]: I1205 10:45:54.975416 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-kubelet-dir\") pod \"installer-1-retry-1-master-0\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:45:54.975776 master-0 kubenswrapper[10444]: I1205 10:45:54.975529 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-var-lock\") pod \"installer-1-retry-1-master-0\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:45:54.975776 master-0 kubenswrapper[10444]: I1205 10:45:54.975540 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-kubelet-dir\") pod \"installer-1-retry-1-master-0\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:45:55.478583 master-0 kubenswrapper[10444]: I1205 10:45:55.478505 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d1a50dc7-2abb-4183-a391-6f75ab05110c-kube-api-access\") pod \"installer-1-retry-1-master-0\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:45:55.558841 master-0 kubenswrapper[10444]: I1205 10:45:55.558789 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:55.558841 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:55.558841 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:55.558841 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:55.559113 master-0 kubenswrapper[10444]: I1205 10:45:55.558854 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:55.574666 master-0 kubenswrapper[10444]: I1205 10:45:55.574636 10444 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["kube-system/bootstrap-kube-scheduler-master-0"] Dec 05 10:45:55.575078 master-0 kubenswrapper[10444]: I1205 10:45:55.574818 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="kube-system/bootstrap-kube-scheduler-master-0" podUID="5e09e2af7200e6f9be469dbfd9bb1127" containerName="kube-scheduler" containerID="cri-o://2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000" gracePeriod=30 Dec 05 10:45:55.576828 master-0 kubenswrapper[10444]: I1205 10:45:55.576803 10444 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-scheduler/openshift-kube-scheduler-master-0"] Dec 05 10:45:55.577085 master-0 kubenswrapper[10444]: E1205 10:45:55.577055 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e09e2af7200e6f9be469dbfd9bb1127" containerName="kube-scheduler" Dec 05 10:45:55.577085 master-0 kubenswrapper[10444]: I1205 10:45:55.577072 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e09e2af7200e6f9be469dbfd9bb1127" containerName="kube-scheduler" Dec 05 10:45:55.577085 master-0 kubenswrapper[10444]: E1205 10:45:55.577084 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e09e2af7200e6f9be469dbfd9bb1127" containerName="kube-scheduler" Dec 05 10:45:55.577225 master-0 kubenswrapper[10444]: I1205 10:45:55.577091 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e09e2af7200e6f9be469dbfd9bb1127" containerName="kube-scheduler" Dec 05 10:45:55.577225 master-0 kubenswrapper[10444]: I1205 10:45:55.577196 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e09e2af7200e6f9be469dbfd9bb1127" containerName="kube-scheduler" Dec 05 10:45:55.577225 master-0 kubenswrapper[10444]: I1205 10:45:55.577210 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e09e2af7200e6f9be469dbfd9bb1127" containerName="kube-scheduler" Dec 05 10:45:55.578236 master-0 kubenswrapper[10444]: I1205 10:45:55.578202 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:45:55.640703 master-0 kubenswrapper[10444]: I1205 10:45:55.640587 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:45:55.684913 master-0 kubenswrapper[10444]: I1205 10:45:55.684849 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-resource-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:45:55.684913 master-0 kubenswrapper[10444]: I1205 10:45:55.684901 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-cert-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:45:55.786312 master-0 kubenswrapper[10444]: I1205 10:45:55.786219 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-resource-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:45:55.786573 master-0 kubenswrapper[10444]: I1205 10:45:55.786455 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-cert-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:45:55.786573 master-0 kubenswrapper[10444]: I1205 10:45:55.786455 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-resource-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:45:55.786662 master-0 kubenswrapper[10444]: I1205 10:45:55.786596 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-cert-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:45:55.932588 master-0 kubenswrapper[10444]: I1205 10:45:55.931317 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-m42rr"] Dec 05 10:45:55.932588 master-0 kubenswrapper[10444]: I1205 10:45:55.931588 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" podUID="c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" gracePeriod=30 Dec 05 10:45:55.954715 master-0 kubenswrapper[10444]: I1205 10:45:55.954638 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:45:55.965531 master-0 kubenswrapper[10444]: I1205 10:45:55.965288 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-master-0"] Dec 05 10:45:56.269944 master-0 kubenswrapper[10444]: I1205 10:45:56.269888 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:45:56.288823 master-0 kubenswrapper[10444]: I1205 10:45:56.288746 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-1-retry-1-master-0"] Dec 05 10:45:56.290555 master-0 kubenswrapper[10444]: I1205 10:45:56.290496 10444 generic.go:334] "Generic (PLEG): container finished" podID="b3c5922c-65b0-4c75-992c-f9c88f81762d" containerID="f02bbe84a0ab6b5c3f31276301465d91b619a21b309cbdf880fc6c1df1dcf605" exitCode=0 Dec 05 10:45:56.290643 master-0 kubenswrapper[10444]: I1205 10:45:56.290577 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-4-master-0" event={"ID":"b3c5922c-65b0-4c75-992c-f9c88f81762d","Type":"ContainerDied","Data":"f02bbe84a0ab6b5c3f31276301465d91b619a21b309cbdf880fc6c1df1dcf605"} Dec 05 10:45:56.292444 master-0 kubenswrapper[10444]: I1205 10:45:56.292402 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerStarted","Data":"b73ea430e90c45d058f4dc00766a5e44931feca970feb8976e3fca4580bc112f"} Dec 05 10:45:56.295662 master-0 kubenswrapper[10444]: I1205 10:45:56.295623 10444 generic.go:334] "Generic (PLEG): container finished" podID="5e09e2af7200e6f9be469dbfd9bb1127" containerID="2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000" exitCode=0 Dec 05 10:45:56.295745 master-0 kubenswrapper[10444]: I1205 10:45:56.295685 10444 scope.go:117] "RemoveContainer" containerID="2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000" Dec 05 10:45:56.295867 master-0 kubenswrapper[10444]: I1205 10:45:56.295843 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="kube-system/bootstrap-kube-scheduler-master-0" Dec 05 10:45:56.296467 master-0 kubenswrapper[10444]: W1205 10:45:56.296409 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podd1a50dc7_2abb_4183_a391_6f75ab05110c.slice/crio-f732d3bb1f43e98701b0b37e40366da756058ab74ad9bf8a27a08f013eb33060 WatchSource:0}: Error finding container f732d3bb1f43e98701b0b37e40366da756058ab74ad9bf8a27a08f013eb33060: Status 404 returned error can't find the container with id f732d3bb1f43e98701b0b37e40366da756058ab74ad9bf8a27a08f013eb33060 Dec 05 10:45:56.344794 master-0 kubenswrapper[10444]: I1205 10:45:56.344600 10444 scope.go:117] "RemoveContainer" containerID="f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782" Dec 05 10:45:56.384480 master-0 kubenswrapper[10444]: I1205 10:45:56.384431 10444 scope.go:117] "RemoveContainer" containerID="2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000" Dec 05 10:45:56.385591 master-0 kubenswrapper[10444]: E1205 10:45:56.384935 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000\": container with ID starting with 2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000 not found: ID does not exist" containerID="2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000" Dec 05 10:45:56.385591 master-0 kubenswrapper[10444]: I1205 10:45:56.384999 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000"} err="failed to get container status \"2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000\": rpc error: code = NotFound desc = could not find container \"2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000\": container with ID starting with 2802349f3c37913962bf76abd70cc068ced790cc5227082ed50e6322ea462000 not found: ID does not exist" Dec 05 10:45:56.385591 master-0 kubenswrapper[10444]: I1205 10:45:56.385025 10444 scope.go:117] "RemoveContainer" containerID="f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782" Dec 05 10:45:56.385591 master-0 kubenswrapper[10444]: E1205 10:45:56.385478 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782\": container with ID starting with f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782 not found: ID does not exist" containerID="f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782" Dec 05 10:45:56.385591 master-0 kubenswrapper[10444]: I1205 10:45:56.385505 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782"} err="failed to get container status \"f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782\": rpc error: code = NotFound desc = could not find container \"f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782\": container with ID starting with f776aeb301b1d6208bd3c17cd609758f324f2d8cdb991d8f696e67b96ec0b782 not found: ID does not exist" Dec 05 10:45:56.394009 master-0 kubenswrapper[10444]: I1205 10:45:56.393939 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-logs\") pod \"5e09e2af7200e6f9be469dbfd9bb1127\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " Dec 05 10:45:56.394166 master-0 kubenswrapper[10444]: I1205 10:45:56.394060 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-secrets\") pod \"5e09e2af7200e6f9be469dbfd9bb1127\" (UID: \"5e09e2af7200e6f9be469dbfd9bb1127\") " Dec 05 10:45:56.394245 master-0 kubenswrapper[10444]: I1205 10:45:56.394204 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-secrets" (OuterVolumeSpecName: "secrets") pod "5e09e2af7200e6f9be469dbfd9bb1127" (UID: "5e09e2af7200e6f9be469dbfd9bb1127"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:45:56.394245 master-0 kubenswrapper[10444]: I1205 10:45:56.394198 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-logs" (OuterVolumeSpecName: "logs") pod "5e09e2af7200e6f9be469dbfd9bb1127" (UID: "5e09e2af7200e6f9be469dbfd9bb1127"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:45:56.394494 master-0 kubenswrapper[10444]: I1205 10:45:56.394472 10444 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 10:45:56.394641 master-0 kubenswrapper[10444]: I1205 10:45:56.394495 10444 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/5e09e2af7200e6f9be469dbfd9bb1127-secrets\") on node \"master-0\" DevicePath \"\"" Dec 05 10:45:56.558283 master-0 kubenswrapper[10444]: I1205 10:45:56.558204 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:56.558283 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:56.558283 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:56.558283 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:56.558560 master-0 kubenswrapper[10444]: I1205 10:45:56.558291 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:57.305374 master-0 kubenswrapper[10444]: I1205 10:45:57.305275 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-1-retry-1-master-0" event={"ID":"d1a50dc7-2abb-4183-a391-6f75ab05110c","Type":"ContainerStarted","Data":"f732d3bb1f43e98701b0b37e40366da756058ab74ad9bf8a27a08f013eb33060"} Dec 05 10:45:57.308018 master-0 kubenswrapper[10444]: I1205 10:45:57.307955 10444 generic.go:334] "Generic (PLEG): container finished" podID="bdd20c68f1ccfd1240fb273af0638d01" containerID="baa0fb92990e2611d666dd15d06398cf84d61306c4b64819cbec0b7a2dd1f1f8" exitCode=0 Dec 05 10:45:57.308222 master-0 kubenswrapper[10444]: I1205 10:45:57.308032 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerDied","Data":"baa0fb92990e2611d666dd15d06398cf84d61306c4b64819cbec0b7a2dd1f1f8"} Dec 05 10:45:57.559460 master-0 kubenswrapper[10444]: I1205 10:45:57.559304 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:57.559460 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:57.559460 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:57.559460 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:57.559460 master-0 kubenswrapper[10444]: I1205 10:45:57.559366 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:57.611827 master-0 kubenswrapper[10444]: I1205 10:45:57.611772 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e09e2af7200e6f9be469dbfd9bb1127" path="/var/lib/kubelet/pods/5e09e2af7200e6f9be469dbfd9bb1127/volumes" Dec 05 10:45:57.612143 master-0 kubenswrapper[10444]: I1205 10:45:57.612116 10444 mirror_client.go:130] "Deleting a mirror pod" pod="kube-system/bootstrap-kube-scheduler-master-0" podUID="" Dec 05 10:45:57.639691 master-0 kubenswrapper[10444]: I1205 10:45:57.639598 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:57.818639 master-0 kubenswrapper[10444]: I1205 10:45:57.818414 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-kubelet-dir\") pod \"b3c5922c-65b0-4c75-992c-f9c88f81762d\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " Dec 05 10:45:57.818639 master-0 kubenswrapper[10444]: I1205 10:45:57.818555 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b3c5922c-65b0-4c75-992c-f9c88f81762d" (UID: "b3c5922c-65b0-4c75-992c-f9c88f81762d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:45:57.818994 master-0 kubenswrapper[10444]: I1205 10:45:57.818700 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-var-lock\") pod \"b3c5922c-65b0-4c75-992c-f9c88f81762d\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " Dec 05 10:45:57.818994 master-0 kubenswrapper[10444]: I1205 10:45:57.818758 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-var-lock" (OuterVolumeSpecName: "var-lock") pod "b3c5922c-65b0-4c75-992c-f9c88f81762d" (UID: "b3c5922c-65b0-4c75-992c-f9c88f81762d"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:45:57.818994 master-0 kubenswrapper[10444]: I1205 10:45:57.818767 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b3c5922c-65b0-4c75-992c-f9c88f81762d-kube-api-access\") pod \"b3c5922c-65b0-4c75-992c-f9c88f81762d\" (UID: \"b3c5922c-65b0-4c75-992c-f9c88f81762d\") " Dec 05 10:45:57.819635 master-0 kubenswrapper[10444]: I1205 10:45:57.819563 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:45:57.819635 master-0 kubenswrapper[10444]: I1205 10:45:57.819618 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b3c5922c-65b0-4c75-992c-f9c88f81762d-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:45:57.821839 master-0 kubenswrapper[10444]: I1205 10:45:57.821717 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3c5922c-65b0-4c75-992c-f9c88f81762d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b3c5922c-65b0-4c75-992c-f9c88f81762d" (UID: "b3c5922c-65b0-4c75-992c-f9c88f81762d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:45:57.921200 master-0 kubenswrapper[10444]: I1205 10:45:57.921083 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b3c5922c-65b0-4c75-992c-f9c88f81762d-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:45:58.318056 master-0 kubenswrapper[10444]: I1205 10:45:58.318009 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:45:58.559630 master-0 kubenswrapper[10444]: I1205 10:45:58.559515 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:58.559630 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:58.559630 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:58.559630 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:58.559630 master-0 kubenswrapper[10444]: I1205 10:45:58.559627 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:59.559565 master-0 kubenswrapper[10444]: I1205 10:45:59.559413 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:45:59.559565 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:45:59.559565 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:45:59.559565 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:45:59.559565 master-0 kubenswrapper[10444]: I1205 10:45:59.559553 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:45:59.665113 master-0 kubenswrapper[10444]: E1205 10:45:59.665009 10444 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 05 10:45:59.667124 master-0 kubenswrapper[10444]: E1205 10:45:59.667063 10444 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 05 10:45:59.669227 master-0 kubenswrapper[10444]: E1205 10:45:59.669144 10444 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 05 10:45:59.669366 master-0 kubenswrapper[10444]: E1205 10:45:59.669242 10444 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" podUID="c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" containerName="kube-multus-additional-cni-plugins" Dec 05 10:46:00.521391 master-0 kubenswrapper[10444]: E1205 10:46:00.521301 10444 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.928s" Dec 05 10:46:00.521391 master-0 kubenswrapper[10444]: I1205 10:46:00.521373 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["kube-system/bootstrap-kube-scheduler-master-0"] Dec 05 10:46:00.521391 master-0 kubenswrapper[10444]: I1205 10:46:00.521397 10444 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="kube-system/bootstrap-kube-scheduler-master-0" mirrorPodUID="ea149711-9799-4257-ba5c-349198b203f3" Dec 05 10:46:00.521828 master-0 kubenswrapper[10444]: I1205 10:46:00.521476 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/installer-1-retry-1-master-0"] Dec 05 10:46:00.521828 master-0 kubenswrapper[10444]: I1205 10:46:00.521552 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-1-retry-1-master-0" event={"ID":"d1a50dc7-2abb-4183-a391-6f75ab05110c","Type":"ContainerStarted","Data":"547e7f664a38b79354f91cf2e7d99ad98e5bf34580f28bf7a6e34c034b25b010"} Dec 05 10:46:00.521828 master-0 kubenswrapper[10444]: I1205 10:46:00.521580 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/installer-4-master-0" event={"ID":"b3c5922c-65b0-4c75-992c-f9c88f81762d","Type":"ContainerDied","Data":"4293a095feecc2a652807fec9aa96d59e590e18262f101aef59da2ccb3b4fbf6"} Dec 05 10:46:00.521828 master-0 kubenswrapper[10444]: I1205 10:46:00.521601 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4293a095feecc2a652807fec9aa96d59e590e18262f101aef59da2ccb3b4fbf6" Dec 05 10:46:00.560772 master-0 kubenswrapper[10444]: I1205 10:46:00.560725 10444 mirror_client.go:130] "Deleting a mirror pod" pod="kube-system/bootstrap-kube-scheduler-master-0" podUID="" Dec 05 10:46:00.562273 master-0 kubenswrapper[10444]: I1205 10:46:00.562246 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:00.562273 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:00.562273 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:00.562273 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:00.562416 master-0 kubenswrapper[10444]: I1205 10:46:00.562283 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:00.595611 master-0 kubenswrapper[10444]: I1205 10:46:00.595547 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["kube-system/bootstrap-kube-scheduler-master-0"] Dec 05 10:46:00.595611 master-0 kubenswrapper[10444]: I1205 10:46:00.595595 10444 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="kube-system/bootstrap-kube-scheduler-master-0" mirrorPodUID="ea149711-9799-4257-ba5c-349198b203f3" Dec 05 10:46:01.341636 master-0 kubenswrapper[10444]: I1205 10:46:01.341553 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerStarted","Data":"29ac63e75838b26e5ffeb47eea93679b6f563ae1d361a9d774732d46e7d9c13d"} Dec 05 10:46:01.341636 master-0 kubenswrapper[10444]: I1205 10:46:01.341609 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerStarted","Data":"804cb8acb4e8cb78d7a134f3b11dcb55d43d2e3b4630654676497a1d400dbc6f"} Dec 05 10:46:01.341636 master-0 kubenswrapper[10444]: I1205 10:46:01.341622 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerStarted","Data":"8c491ff443ec7786bd22bb8e647eeff7f1b6e812a627f82eec65955c76037686"} Dec 05 10:46:01.341977 master-0 kubenswrapper[10444]: I1205 10:46:01.341805 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:46:01.558825 master-0 kubenswrapper[10444]: I1205 10:46:01.558760 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:01.558825 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:01.558825 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:01.558825 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:01.559138 master-0 kubenswrapper[10444]: I1205 10:46:01.558832 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:01.672930 master-0 kubenswrapper[10444]: I1205 10:46:01.672772 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-1-retry-1-master-0" podStartSLOduration=7.672754188 podStartE2EDuration="7.672754188s" podCreationTimestamp="2025-12-05 10:45:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:46:01.670222315 +0000 UTC m=+502.262134902" watchObservedRunningTime="2025-12-05 10:46:01.672754188 +0000 UTC m=+502.264666755" Dec 05 10:46:01.691117 master-0 kubenswrapper[10444]: I1205 10:46:01.691046 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" podStartSLOduration=6.691031015 podStartE2EDuration="6.691031015s" podCreationTimestamp="2025-12-05 10:45:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:46:01.690357459 +0000 UTC m=+502.282270046" watchObservedRunningTime="2025-12-05 10:46:01.691031015 +0000 UTC m=+502.282943582" Dec 05 10:46:02.351017 master-0 kubenswrapper[10444]: I1205 10:46:02.350935 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/installer-1-retry-1-master-0" podUID="d1a50dc7-2abb-4183-a391-6f75ab05110c" containerName="installer" containerID="cri-o://547e7f664a38b79354f91cf2e7d99ad98e5bf34580f28bf7a6e34c034b25b010" gracePeriod=30 Dec 05 10:46:02.559591 master-0 kubenswrapper[10444]: I1205 10:46:02.559485 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:02.559591 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:02.559591 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:02.559591 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:02.560026 master-0 kubenswrapper[10444]: I1205 10:46:02.559614 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:03.558608 master-0 kubenswrapper[10444]: I1205 10:46:03.558547 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:03.558608 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:03.558608 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:03.558608 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:03.559161 master-0 kubenswrapper[10444]: I1205 10:46:03.558615 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:04.558950 master-0 kubenswrapper[10444]: I1205 10:46:04.558870 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:04.558950 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:04.558950 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:04.558950 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:04.559537 master-0 kubenswrapper[10444]: I1205 10:46:04.558956 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:05.559702 master-0 kubenswrapper[10444]: I1205 10:46:05.559608 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:05.559702 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:05.559702 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:05.559702 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:05.560334 master-0 kubenswrapper[10444]: I1205 10:46:05.559721 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:05.702934 master-0 kubenswrapper[10444]: I1205 10:46:05.702849 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-2-master-0"] Dec 05 10:46:05.703388 master-0 kubenswrapper[10444]: E1205 10:46:05.703347 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c5922c-65b0-4c75-992c-f9c88f81762d" containerName="installer" Dec 05 10:46:05.703388 master-0 kubenswrapper[10444]: I1205 10:46:05.703385 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c5922c-65b0-4c75-992c-f9c88f81762d" containerName="installer" Dec 05 10:46:05.703675 master-0 kubenswrapper[10444]: I1205 10:46:05.703644 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3c5922c-65b0-4c75-992c-f9c88f81762d" containerName="installer" Dec 05 10:46:05.704356 master-0 kubenswrapper[10444]: I1205 10:46:05.704317 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:05.715992 master-0 kubenswrapper[10444]: I1205 10:46:05.715917 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-2-master-0"] Dec 05 10:46:05.740707 master-0 kubenswrapper[10444]: I1205 10:46:05.740403 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-var-lock\") pod \"installer-2-master-0\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:05.740707 master-0 kubenswrapper[10444]: I1205 10:46:05.740667 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:05.740707 master-0 kubenswrapper[10444]: I1205 10:46:05.740704 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kube-api-access\") pod \"installer-2-master-0\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:05.841517 master-0 kubenswrapper[10444]: I1205 10:46:05.841319 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-var-lock\") pod \"installer-2-master-0\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:05.841517 master-0 kubenswrapper[10444]: I1205 10:46:05.841386 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:05.841517 master-0 kubenswrapper[10444]: I1205 10:46:05.841409 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kube-api-access\") pod \"installer-2-master-0\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:05.841517 master-0 kubenswrapper[10444]: I1205 10:46:05.841492 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-var-lock\") pod \"installer-2-master-0\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:05.841942 master-0 kubenswrapper[10444]: I1205 10:46:05.841569 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:05.858503 master-0 kubenswrapper[10444]: I1205 10:46:05.858436 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kube-api-access\") pod \"installer-2-master-0\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:06.032284 master-0 kubenswrapper[10444]: I1205 10:46:06.032205 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:06.344754 master-0 kubenswrapper[10444]: I1205 10:46:06.344683 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-8dbbb5754-7p9c2"] Dec 05 10:46:06.346051 master-0 kubenswrapper[10444]: I1205 10:46:06.345966 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:46:06.348614 master-0 kubenswrapper[10444]: I1205 10:46:06.348555 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-5972r" Dec 05 10:46:06.357444 master-0 kubenswrapper[10444]: I1205 10:46:06.357377 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-8dbbb5754-7p9c2"] Dec 05 10:46:06.468443 master-0 kubenswrapper[10444]: I1205 10:46:06.468263 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gfgm\" (UniqueName: \"kubernetes.io/projected/3332b604-6f0d-4243-bce1-b3ab76e58c78-kube-api-access-6gfgm\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:46:06.468657 master-0 kubenswrapper[10444]: I1205 10:46:06.468499 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:46:06.512110 master-0 kubenswrapper[10444]: W1205 10:46:06.512010 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podb18ccfcf_5f9a_4f79_a919_2e7819801a53.slice/crio-89e079fe893703438f560afed93460e02bf0749f65d41a362ae1c69d6bb9fc0d WatchSource:0}: Error finding container 89e079fe893703438f560afed93460e02bf0749f65d41a362ae1c69d6bb9fc0d: Status 404 returned error can't find the container with id 89e079fe893703438f560afed93460e02bf0749f65d41a362ae1c69d6bb9fc0d Dec 05 10:46:06.512365 master-0 kubenswrapper[10444]: I1205 10:46:06.512094 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-2-master-0"] Dec 05 10:46:06.558471 master-0 kubenswrapper[10444]: I1205 10:46:06.558386 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:06.558471 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:06.558471 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:06.558471 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:06.558807 master-0 kubenswrapper[10444]: I1205 10:46:06.558502 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:06.569695 master-0 kubenswrapper[10444]: I1205 10:46:06.569648 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:46:06.570138 master-0 kubenswrapper[10444]: I1205 10:46:06.569860 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gfgm\" (UniqueName: \"kubernetes.io/projected/3332b604-6f0d-4243-bce1-b3ab76e58c78-kube-api-access-6gfgm\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:46:06.577047 master-0 kubenswrapper[10444]: I1205 10:46:06.576983 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:46:06.595055 master-0 kubenswrapper[10444]: I1205 10:46:06.595012 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gfgm\" (UniqueName: \"kubernetes.io/projected/3332b604-6f0d-4243-bce1-b3ab76e58c78-kube-api-access-6gfgm\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:46:06.670265 master-0 kubenswrapper[10444]: I1205 10:46:06.670220 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:46:07.077870 master-0 kubenswrapper[10444]: I1205 10:46:07.077798 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-8dbbb5754-7p9c2"] Dec 05 10:46:07.085800 master-0 kubenswrapper[10444]: W1205 10:46:07.085737 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3332b604_6f0d_4243_bce1_b3ab76e58c78.slice/crio-2543decbaf3eadc4ac21f8d8b5c5c326c025c08f080ed2e80c25c76a6e2b34dd WatchSource:0}: Error finding container 2543decbaf3eadc4ac21f8d8b5c5c326c025c08f080ed2e80c25c76a6e2b34dd: Status 404 returned error can't find the container with id 2543decbaf3eadc4ac21f8d8b5c5c326c025c08f080ed2e80c25c76a6e2b34dd Dec 05 10:46:07.387320 master-0 kubenswrapper[10444]: I1205 10:46:07.387268 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-2-master-0" event={"ID":"b18ccfcf-5f9a-4f79-a919-2e7819801a53","Type":"ContainerStarted","Data":"86a6b1fad7c7b87e10e7f7445ee24f3a149b2ea3314654619f1a4096f0680d4f"} Dec 05 10:46:07.387320 master-0 kubenswrapper[10444]: I1205 10:46:07.387321 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-2-master-0" event={"ID":"b18ccfcf-5f9a-4f79-a919-2e7819801a53","Type":"ContainerStarted","Data":"89e079fe893703438f560afed93460e02bf0749f65d41a362ae1c69d6bb9fc0d"} Dec 05 10:46:07.390256 master-0 kubenswrapper[10444]: I1205 10:46:07.390215 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" event={"ID":"3332b604-6f0d-4243-bce1-b3ab76e58c78","Type":"ContainerStarted","Data":"46c395eca9c065a071d9bf2865f62933c63d98c9c02d5c3bf0fc95b97a4d1526"} Dec 05 10:46:07.390256 master-0 kubenswrapper[10444]: I1205 10:46:07.390259 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" event={"ID":"3332b604-6f0d-4243-bce1-b3ab76e58c78","Type":"ContainerStarted","Data":"2543decbaf3eadc4ac21f8d8b5c5c326c025c08f080ed2e80c25c76a6e2b34dd"} Dec 05 10:46:07.406676 master-0 kubenswrapper[10444]: I1205 10:46:07.406602 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-2-master-0" podStartSLOduration=2.406583693 podStartE2EDuration="2.406583693s" podCreationTimestamp="2025-12-05 10:46:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:46:07.405596689 +0000 UTC m=+507.997509276" watchObservedRunningTime="2025-12-05 10:46:07.406583693 +0000 UTC m=+507.998496260" Dec 05 10:46:07.558677 master-0 kubenswrapper[10444]: I1205 10:46:07.558532 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:07.558677 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:07.558677 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:07.558677 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:07.558677 master-0 kubenswrapper[10444]: I1205 10:46:07.558580 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:08.400159 master-0 kubenswrapper[10444]: I1205 10:46:08.400100 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" event={"ID":"3332b604-6f0d-4243-bce1-b3ab76e58c78","Type":"ContainerStarted","Data":"6467221d244ea62ac0efcea600a5fcec4b4fe68b5d8561bb24fd6a84a9d02b81"} Dec 05 10:46:08.417573 master-0 kubenswrapper[10444]: I1205 10:46:08.417491 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" podStartSLOduration=2.417475808 podStartE2EDuration="2.417475808s" podCreationTimestamp="2025-12-05 10:46:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:46:08.416173125 +0000 UTC m=+509.008085702" watchObservedRunningTime="2025-12-05 10:46:08.417475808 +0000 UTC m=+509.009388365" Dec 05 10:46:08.477205 master-0 kubenswrapper[10444]: I1205 10:46:08.476301 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/multus-admission-controller-7dfc5b745f-67rx7"] Dec 05 10:46:08.477205 master-0 kubenswrapper[10444]: I1205 10:46:08.476579 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" podUID="8d76404b-6d62-4a61-b6f6-0c8073eba198" containerName="multus-admission-controller" containerID="cri-o://163ef91a3ca09ff8a80da2c166f293609027eb6a5251a7136d985bf676587d9d" gracePeriod=30 Dec 05 10:46:08.477205 master-0 kubenswrapper[10444]: I1205 10:46:08.476965 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" podUID="8d76404b-6d62-4a61-b6f6-0c8073eba198" containerName="kube-rbac-proxy" containerID="cri-o://d70a4de37cbf6d2709586eaecfa831636df1b7ccf32f015e3097c3c53b817cac" gracePeriod=30 Dec 05 10:46:08.558582 master-0 kubenswrapper[10444]: I1205 10:46:08.558540 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:08.558582 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:08.558582 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:08.558582 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:08.558834 master-0 kubenswrapper[10444]: I1205 10:46:08.558607 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:09.413470 master-0 kubenswrapper[10444]: I1205 10:46:09.413338 10444 generic.go:334] "Generic (PLEG): container finished" podID="8d76404b-6d62-4a61-b6f6-0c8073eba198" containerID="d70a4de37cbf6d2709586eaecfa831636df1b7ccf32f015e3097c3c53b817cac" exitCode=0 Dec 05 10:46:09.413470 master-0 kubenswrapper[10444]: I1205 10:46:09.413405 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" event={"ID":"8d76404b-6d62-4a61-b6f6-0c8073eba198","Type":"ContainerDied","Data":"d70a4de37cbf6d2709586eaecfa831636df1b7ccf32f015e3097c3c53b817cac"} Dec 05 10:46:09.560013 master-0 kubenswrapper[10444]: I1205 10:46:09.559937 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:09.560013 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:09.560013 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:09.560013 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:09.560297 master-0 kubenswrapper[10444]: I1205 10:46:09.560055 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:09.664758 master-0 kubenswrapper[10444]: E1205 10:46:09.664590 10444 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 05 10:46:09.666275 master-0 kubenswrapper[10444]: E1205 10:46:09.666197 10444 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 05 10:46:09.667700 master-0 kubenswrapper[10444]: E1205 10:46:09.667652 10444 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 05 10:46:09.667789 master-0 kubenswrapper[10444]: E1205 10:46:09.667697 10444 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" podUID="c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" containerName="kube-multus-additional-cni-plugins" Dec 05 10:46:10.559704 master-0 kubenswrapper[10444]: I1205 10:46:10.559396 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:10.559704 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:10.559704 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:10.559704 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:10.559704 master-0 kubenswrapper[10444]: I1205 10:46:10.559495 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:11.559795 master-0 kubenswrapper[10444]: I1205 10:46:11.559733 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:11.559795 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:11.559795 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:11.559795 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:11.560472 master-0 kubenswrapper[10444]: I1205 10:46:11.559798 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:12.558903 master-0 kubenswrapper[10444]: I1205 10:46:12.558809 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:12.558903 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:12.558903 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:12.558903 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:12.558903 master-0 kubenswrapper[10444]: I1205 10:46:12.558897 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:13.558551 master-0 kubenswrapper[10444]: I1205 10:46:13.558501 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:13.558551 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:13.558551 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:13.558551 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:13.559164 master-0 kubenswrapper[10444]: I1205 10:46:13.558571 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:14.558992 master-0 kubenswrapper[10444]: I1205 10:46:14.558935 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:14.558992 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:14.558992 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:14.558992 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:14.559550 master-0 kubenswrapper[10444]: I1205 10:46:14.559004 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:15.559183 master-0 kubenswrapper[10444]: I1205 10:46:15.559108 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:15.559183 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:15.559183 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:15.559183 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:15.559183 master-0 kubenswrapper[10444]: I1205 10:46:15.559174 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:16.559746 master-0 kubenswrapper[10444]: I1205 10:46:16.559657 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:16.559746 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:16.559746 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:16.559746 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:16.560648 master-0 kubenswrapper[10444]: I1205 10:46:16.559757 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:17.558968 master-0 kubenswrapper[10444]: I1205 10:46:17.558900 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:17.558968 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:17.558968 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:17.558968 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:17.559332 master-0 kubenswrapper[10444]: I1205 10:46:17.559003 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:18.558628 master-0 kubenswrapper[10444]: I1205 10:46:18.558539 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:18.558628 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:18.558628 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:18.558628 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:18.558628 master-0 kubenswrapper[10444]: I1205 10:46:18.558611 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:19.558098 master-0 kubenswrapper[10444]: I1205 10:46:19.557998 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:19.558098 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:19.558098 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:19.558098 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:19.558098 master-0 kubenswrapper[10444]: I1205 10:46:19.558062 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:19.663743 master-0 kubenswrapper[10444]: E1205 10:46:19.663629 10444 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 05 10:46:19.665721 master-0 kubenswrapper[10444]: E1205 10:46:19.665662 10444 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 05 10:46:19.667514 master-0 kubenswrapper[10444]: E1205 10:46:19.667469 10444 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 05 10:46:19.667575 master-0 kubenswrapper[10444]: E1205 10:46:19.667515 10444 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" podUID="c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" containerName="kube-multus-additional-cni-plugins" Dec 05 10:46:20.559189 master-0 kubenswrapper[10444]: I1205 10:46:20.559084 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:20.559189 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:20.559189 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:20.559189 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:20.559503 master-0 kubenswrapper[10444]: I1205 10:46:20.559210 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:21.559247 master-0 kubenswrapper[10444]: I1205 10:46:21.559169 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:21.559247 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:21.559247 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:21.559247 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:21.559894 master-0 kubenswrapper[10444]: I1205 10:46:21.559260 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:22.322503 master-0 kubenswrapper[10444]: I1205 10:46:22.321506 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/installer-2-master-0"] Dec 05 10:46:22.322503 master-0 kubenswrapper[10444]: I1205 10:46:22.321804 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/installer-2-master-0" podUID="b18ccfcf-5f9a-4f79-a919-2e7819801a53" containerName="installer" containerID="cri-o://86a6b1fad7c7b87e10e7f7445ee24f3a149b2ea3314654619f1a4096f0680d4f" gracePeriod=30 Dec 05 10:46:22.527683 master-0 kubenswrapper[10444]: I1205 10:46:22.527615 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-2-master-0_b18ccfcf-5f9a-4f79-a919-2e7819801a53/installer/0.log" Dec 05 10:46:22.527891 master-0 kubenswrapper[10444]: I1205 10:46:22.527734 10444 generic.go:334] "Generic (PLEG): container finished" podID="b18ccfcf-5f9a-4f79-a919-2e7819801a53" containerID="86a6b1fad7c7b87e10e7f7445ee24f3a149b2ea3314654619f1a4096f0680d4f" exitCode=1 Dec 05 10:46:22.527891 master-0 kubenswrapper[10444]: I1205 10:46:22.527809 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-2-master-0" event={"ID":"b18ccfcf-5f9a-4f79-a919-2e7819801a53","Type":"ContainerDied","Data":"86a6b1fad7c7b87e10e7f7445ee24f3a149b2ea3314654619f1a4096f0680d4f"} Dec 05 10:46:22.558982 master-0 kubenswrapper[10444]: I1205 10:46:22.558922 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:22.558982 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:22.558982 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:22.558982 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:22.559172 master-0 kubenswrapper[10444]: I1205 10:46:22.558993 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:22.770560 master-0 kubenswrapper[10444]: I1205 10:46:22.770118 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-2-master-0_b18ccfcf-5f9a-4f79-a919-2e7819801a53/installer/0.log" Dec 05 10:46:22.771157 master-0 kubenswrapper[10444]: I1205 10:46:22.770625 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:22.820286 master-0 kubenswrapper[10444]: I1205 10:46:22.820211 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kubelet-dir\") pod \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " Dec 05 10:46:22.820541 master-0 kubenswrapper[10444]: I1205 10:46:22.820310 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-var-lock\") pod \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " Dec 05 10:46:22.820541 master-0 kubenswrapper[10444]: I1205 10:46:22.820386 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b18ccfcf-5f9a-4f79-a919-2e7819801a53" (UID: "b18ccfcf-5f9a-4f79-a919-2e7819801a53"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:46:22.820541 master-0 kubenswrapper[10444]: I1205 10:46:22.820396 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-var-lock" (OuterVolumeSpecName: "var-lock") pod "b18ccfcf-5f9a-4f79-a919-2e7819801a53" (UID: "b18ccfcf-5f9a-4f79-a919-2e7819801a53"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:46:22.820541 master-0 kubenswrapper[10444]: I1205 10:46:22.820420 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kube-api-access\") pod \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\" (UID: \"b18ccfcf-5f9a-4f79-a919-2e7819801a53\") " Dec 05 10:46:22.821056 master-0 kubenswrapper[10444]: I1205 10:46:22.821026 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:22.821056 master-0 kubenswrapper[10444]: I1205 10:46:22.821050 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:22.823088 master-0 kubenswrapper[10444]: I1205 10:46:22.823048 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b18ccfcf-5f9a-4f79-a919-2e7819801a53" (UID: "b18ccfcf-5f9a-4f79-a919-2e7819801a53"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:46:22.922721 master-0 kubenswrapper[10444]: I1205 10:46:22.922514 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b18ccfcf-5f9a-4f79-a919-2e7819801a53-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:23.534756 master-0 kubenswrapper[10444]: I1205 10:46:23.534703 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-2-master-0_b18ccfcf-5f9a-4f79-a919-2e7819801a53/installer/0.log" Dec 05 10:46:23.535009 master-0 kubenswrapper[10444]: I1205 10:46:23.534777 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-2-master-0" event={"ID":"b18ccfcf-5f9a-4f79-a919-2e7819801a53","Type":"ContainerDied","Data":"89e079fe893703438f560afed93460e02bf0749f65d41a362ae1c69d6bb9fc0d"} Dec 05 10:46:23.535009 master-0 kubenswrapper[10444]: I1205 10:46:23.534822 10444 scope.go:117] "RemoveContainer" containerID="86a6b1fad7c7b87e10e7f7445ee24f3a149b2ea3314654619f1a4096f0680d4f" Dec 05 10:46:23.535009 master-0 kubenswrapper[10444]: I1205 10:46:23.534837 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-2-master-0" Dec 05 10:46:23.559902 master-0 kubenswrapper[10444]: I1205 10:46:23.559855 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:23.559902 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:23.559902 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:23.559902 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:23.560080 master-0 kubenswrapper[10444]: I1205 10:46:23.559912 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:23.579126 master-0 kubenswrapper[10444]: I1205 10:46:23.579069 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/installer-2-master-0"] Dec 05 10:46:23.606901 master-0 kubenswrapper[10444]: I1205 10:46:23.606844 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/installer-2-master-0"] Dec 05 10:46:24.559498 master-0 kubenswrapper[10444]: I1205 10:46:24.559378 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:46:24.559498 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:46:24.559498 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:46:24.559498 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:46:24.560181 master-0 kubenswrapper[10444]: I1205 10:46:24.559508 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:46:24.560181 master-0 kubenswrapper[10444]: I1205 10:46:24.559583 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:46:24.560520 master-0 kubenswrapper[10444]: I1205 10:46:24.560476 10444 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="router" containerStatusID={"Type":"cri-o","ID":"df3d2e1cdc52294902ada6e0e872265f937f9cd41ea5e7643f484cd5a4774ce3"} pod="openshift-ingress/router-default-5465c8b4db-s4c2f" containerMessage="Container router failed startup probe, will be restarted" Dec 05 10:46:24.560595 master-0 kubenswrapper[10444]: I1205 10:46:24.560553 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" containerID="cri-o://df3d2e1cdc52294902ada6e0e872265f937f9cd41ea5e7643f484cd5a4774ce3" gracePeriod=3600 Dec 05 10:46:25.604712 master-0 kubenswrapper[10444]: I1205 10:46:25.604633 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b18ccfcf-5f9a-4f79-a919-2e7819801a53" path="/var/lib/kubelet/pods/b18ccfcf-5f9a-4f79-a919-2e7819801a53/volumes" Dec 05 10:46:26.043332 master-0 kubenswrapper[10444]: I1205 10:46:26.043292 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-m42rr_c3034770-d607-4aa2-9b4b-cf1e4b7acf4c/kube-multus-additional-cni-plugins/0.log" Dec 05 10:46:26.043520 master-0 kubenswrapper[10444]: I1205 10:46:26.043358 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:46:26.075001 master-0 kubenswrapper[10444]: I1205 10:46:26.074924 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-cni-sysctl-allowlist\") pod \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " Dec 05 10:46:26.075001 master-0 kubenswrapper[10444]: I1205 10:46:26.075002 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-ready\") pod \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " Dec 05 10:46:26.075277 master-0 kubenswrapper[10444]: I1205 10:46:26.075069 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-tuning-conf-dir\") pod \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " Dec 05 10:46:26.075277 master-0 kubenswrapper[10444]: I1205 10:46:26.075100 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5ld6\" (UniqueName: \"kubernetes.io/projected/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-kube-api-access-s5ld6\") pod \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\" (UID: \"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c\") " Dec 05 10:46:26.075277 master-0 kubenswrapper[10444]: I1205 10:46:26.075161 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" (UID: "c3034770-d607-4aa2-9b4b-cf1e4b7acf4c"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:46:26.075413 master-0 kubenswrapper[10444]: I1205 10:46:26.075351 10444 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-tuning-conf-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:26.075413 master-0 kubenswrapper[10444]: I1205 10:46:26.075375 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" (UID: "c3034770-d607-4aa2-9b4b-cf1e4b7acf4c"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:46:26.075525 master-0 kubenswrapper[10444]: I1205 10:46:26.075474 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-ready" (OuterVolumeSpecName: "ready") pod "c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" (UID: "c3034770-d607-4aa2-9b4b-cf1e4b7acf4c"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:46:26.079707 master-0 kubenswrapper[10444]: I1205 10:46:26.079645 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-kube-api-access-s5ld6" (OuterVolumeSpecName: "kube-api-access-s5ld6") pod "c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" (UID: "c3034770-d607-4aa2-9b4b-cf1e4b7acf4c"). InnerVolumeSpecName "kube-api-access-s5ld6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:46:26.118798 master-0 kubenswrapper[10444]: I1205 10:46:26.118728 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-3-master-0"] Dec 05 10:46:26.119006 master-0 kubenswrapper[10444]: E1205 10:46:26.118997 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b18ccfcf-5f9a-4f79-a919-2e7819801a53" containerName="installer" Dec 05 10:46:26.119042 master-0 kubenswrapper[10444]: I1205 10:46:26.119012 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="b18ccfcf-5f9a-4f79-a919-2e7819801a53" containerName="installer" Dec 05 10:46:26.119042 master-0 kubenswrapper[10444]: E1205 10:46:26.119034 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" containerName="kube-multus-additional-cni-plugins" Dec 05 10:46:26.119042 master-0 kubenswrapper[10444]: I1205 10:46:26.119040 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" containerName="kube-multus-additional-cni-plugins" Dec 05 10:46:26.119202 master-0 kubenswrapper[10444]: I1205 10:46:26.119171 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" containerName="kube-multus-additional-cni-plugins" Dec 05 10:46:26.119242 master-0 kubenswrapper[10444]: I1205 10:46:26.119204 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="b18ccfcf-5f9a-4f79-a919-2e7819801a53" containerName="installer" Dec 05 10:46:26.119692 master-0 kubenswrapper[10444]: I1205 10:46:26.119673 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:46:26.137290 master-0 kubenswrapper[10444]: I1205 10:46:26.137229 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-3-master-0"] Dec 05 10:46:26.175883 master-0 kubenswrapper[10444]: I1205 10:46:26.175742 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:46:26.175883 master-0 kubenswrapper[10444]: I1205 10:46:26.175814 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:46:26.176177 master-0 kubenswrapper[10444]: I1205 10:46:26.175965 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:46:26.176177 master-0 kubenswrapper[10444]: I1205 10:46:26.176147 10444 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-cni-sysctl-allowlist\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:26.176177 master-0 kubenswrapper[10444]: I1205 10:46:26.176170 10444 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-ready\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:26.176302 master-0 kubenswrapper[10444]: I1205 10:46:26.176185 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5ld6\" (UniqueName: \"kubernetes.io/projected/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c-kube-api-access-s5ld6\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:26.277190 master-0 kubenswrapper[10444]: I1205 10:46:26.277028 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:46:26.277393 master-0 kubenswrapper[10444]: I1205 10:46:26.277170 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:46:26.277393 master-0 kubenswrapper[10444]: I1205 10:46:26.277284 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:46:26.277393 master-0 kubenswrapper[10444]: I1205 10:46:26.277365 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:46:26.277393 master-0 kubenswrapper[10444]: I1205 10:46:26.277383 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:46:26.293191 master-0 kubenswrapper[10444]: I1205 10:46:26.293136 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:46:26.446997 master-0 kubenswrapper[10444]: I1205 10:46:26.446935 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:46:26.555628 master-0 kubenswrapper[10444]: I1205 10:46:26.555578 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-m42rr_c3034770-d607-4aa2-9b4b-cf1e4b7acf4c/kube-multus-additional-cni-plugins/0.log" Dec 05 10:46:26.555825 master-0 kubenswrapper[10444]: I1205 10:46:26.555635 10444 generic.go:334] "Generic (PLEG): container finished" podID="c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" exitCode=137 Dec 05 10:46:26.555825 master-0 kubenswrapper[10444]: I1205 10:46:26.555666 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" event={"ID":"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c","Type":"ContainerDied","Data":"97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30"} Dec 05 10:46:26.555825 master-0 kubenswrapper[10444]: I1205 10:46:26.555685 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" Dec 05 10:46:26.555825 master-0 kubenswrapper[10444]: I1205 10:46:26.555709 10444 scope.go:117] "RemoveContainer" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" Dec 05 10:46:26.555825 master-0 kubenswrapper[10444]: I1205 10:46:26.555696 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-m42rr" event={"ID":"c3034770-d607-4aa2-9b4b-cf1e4b7acf4c","Type":"ContainerDied","Data":"bfef724a46fd2289f7233d41b2c5228c9aaaf92dcf1427768b977e1adf40175a"} Dec 05 10:46:26.574484 master-0 kubenswrapper[10444]: I1205 10:46:26.574446 10444 scope.go:117] "RemoveContainer" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" Dec 05 10:46:26.576086 master-0 kubenswrapper[10444]: E1205 10:46:26.576024 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30\": container with ID starting with 97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30 not found: ID does not exist" containerID="97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30" Dec 05 10:46:26.576157 master-0 kubenswrapper[10444]: I1205 10:46:26.576089 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30"} err="failed to get container status \"97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30\": rpc error: code = NotFound desc = could not find container \"97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30\": container with ID starting with 97ea5e542d6ae45c5a07c681a1afba9c84f0842daa39f6d4d165440bfc547b30 not found: ID does not exist" Dec 05 10:46:26.641733 master-0 kubenswrapper[10444]: I1205 10:46:26.641650 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-m42rr"] Dec 05 10:46:26.649056 master-0 kubenswrapper[10444]: I1205 10:46:26.648987 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-m42rr"] Dec 05 10:46:26.912754 master-0 kubenswrapper[10444]: I1205 10:46:26.912689 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-3-master-0"] Dec 05 10:46:27.572737 master-0 kubenswrapper[10444]: I1205 10:46:27.572674 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-3-master-0" event={"ID":"75ba20a1-b322-4ce8-88d2-d8ee210cc27a","Type":"ContainerStarted","Data":"c8f54c8d7593dfbeebbefc04c820ade3fd814ef907ea5e95f2df34b8ed43589a"} Dec 05 10:46:27.574399 master-0 kubenswrapper[10444]: I1205 10:46:27.574361 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-3-master-0" event={"ID":"75ba20a1-b322-4ce8-88d2-d8ee210cc27a","Type":"ContainerStarted","Data":"3b1988a80333a5d160a32b470d3b1aa05cd5259d79366c745b970eff4d6866a0"} Dec 05 10:46:27.603342 master-0 kubenswrapper[10444]: I1205 10:46:27.603262 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-3-master-0" podStartSLOduration=1.6032269860000001 podStartE2EDuration="1.603226986s" podCreationTimestamp="2025-12-05 10:46:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:46:27.601300399 +0000 UTC m=+528.193212976" watchObservedRunningTime="2025-12-05 10:46:27.603226986 +0000 UTC m=+528.195139553" Dec 05 10:46:27.613085 master-0 kubenswrapper[10444]: I1205 10:46:27.613043 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3034770-d607-4aa2-9b4b-cf1e4b7acf4c" path="/var/lib/kubelet/pods/c3034770-d607-4aa2-9b4b-cf1e4b7acf4c/volumes" Dec 05 10:46:30.592026 master-0 kubenswrapper[10444]: I1205 10:46:30.591954 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-1-retry-1-master-0_d1a50dc7-2abb-4183-a391-6f75ab05110c/installer/0.log" Dec 05 10:46:30.592026 master-0 kubenswrapper[10444]: I1205 10:46:30.592031 10444 generic.go:334] "Generic (PLEG): container finished" podID="d1a50dc7-2abb-4183-a391-6f75ab05110c" containerID="547e7f664a38b79354f91cf2e7d99ad98e5bf34580f28bf7a6e34c034b25b010" exitCode=1 Dec 05 10:46:30.592748 master-0 kubenswrapper[10444]: I1205 10:46:30.592068 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-1-retry-1-master-0" event={"ID":"d1a50dc7-2abb-4183-a391-6f75ab05110c","Type":"ContainerDied","Data":"547e7f664a38b79354f91cf2e7d99ad98e5bf34580f28bf7a6e34c034b25b010"} Dec 05 10:46:30.872845 master-0 kubenswrapper[10444]: I1205 10:46:30.872780 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-1-retry-1-master-0_d1a50dc7-2abb-4183-a391-6f75ab05110c/installer/0.log" Dec 05 10:46:30.872845 master-0 kubenswrapper[10444]: I1205 10:46:30.872847 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:46:31.034871 master-0 kubenswrapper[10444]: I1205 10:46:31.034766 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d1a50dc7-2abb-4183-a391-6f75ab05110c-kube-api-access\") pod \"d1a50dc7-2abb-4183-a391-6f75ab05110c\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " Dec 05 10:46:31.035073 master-0 kubenswrapper[10444]: I1205 10:46:31.035050 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-var-lock\") pod \"d1a50dc7-2abb-4183-a391-6f75ab05110c\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " Dec 05 10:46:31.035161 master-0 kubenswrapper[10444]: I1205 10:46:31.035131 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-kubelet-dir\") pod \"d1a50dc7-2abb-4183-a391-6f75ab05110c\" (UID: \"d1a50dc7-2abb-4183-a391-6f75ab05110c\") " Dec 05 10:46:31.035483 master-0 kubenswrapper[10444]: I1205 10:46:31.035288 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d1a50dc7-2abb-4183-a391-6f75ab05110c" (UID: "d1a50dc7-2abb-4183-a391-6f75ab05110c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:46:31.035483 master-0 kubenswrapper[10444]: I1205 10:46:31.035301 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-var-lock" (OuterVolumeSpecName: "var-lock") pod "d1a50dc7-2abb-4183-a391-6f75ab05110c" (UID: "d1a50dc7-2abb-4183-a391-6f75ab05110c"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:46:31.035912 master-0 kubenswrapper[10444]: I1205 10:46:31.035518 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:31.035912 master-0 kubenswrapper[10444]: I1205 10:46:31.035542 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d1a50dc7-2abb-4183-a391-6f75ab05110c-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:31.038318 master-0 kubenswrapper[10444]: I1205 10:46:31.038250 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1a50dc7-2abb-4183-a391-6f75ab05110c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d1a50dc7-2abb-4183-a391-6f75ab05110c" (UID: "d1a50dc7-2abb-4183-a391-6f75ab05110c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:46:31.137208 master-0 kubenswrapper[10444]: I1205 10:46:31.137080 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d1a50dc7-2abb-4183-a391-6f75ab05110c-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:31.603126 master-0 kubenswrapper[10444]: I1205 10:46:31.603065 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-1-retry-1-master-0_d1a50dc7-2abb-4183-a391-6f75ab05110c/installer/0.log" Dec 05 10:46:31.603668 master-0 kubenswrapper[10444]: I1205 10:46:31.603168 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-1-retry-1-master-0" event={"ID":"d1a50dc7-2abb-4183-a391-6f75ab05110c","Type":"ContainerDied","Data":"f732d3bb1f43e98701b0b37e40366da756058ab74ad9bf8a27a08f013eb33060"} Dec 05 10:46:31.603668 master-0 kubenswrapper[10444]: I1205 10:46:31.603222 10444 scope.go:117] "RemoveContainer" containerID="547e7f664a38b79354f91cf2e7d99ad98e5bf34580f28bf7a6e34c034b25b010" Dec 05 10:46:31.603668 master-0 kubenswrapper[10444]: I1205 10:46:31.603253 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-1-retry-1-master-0" Dec 05 10:46:31.637243 master-0 kubenswrapper[10444]: I1205 10:46:31.637183 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/installer-1-retry-1-master-0"] Dec 05 10:46:31.643936 master-0 kubenswrapper[10444]: I1205 10:46:31.643882 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/installer-1-retry-1-master-0"] Dec 05 10:46:33.605382 master-0 kubenswrapper[10444]: I1205 10:46:33.605307 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1a50dc7-2abb-4183-a391-6f75ab05110c" path="/var/lib/kubelet/pods/d1a50dc7-2abb-4183-a391-6f75ab05110c/volumes" Dec 05 10:46:38.670829 master-0 kubenswrapper[10444]: I1205 10:46:38.670748 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-7dfc5b745f-67rx7_8d76404b-6d62-4a61-b6f6-0c8073eba198/multus-admission-controller/0.log" Dec 05 10:46:38.670829 master-0 kubenswrapper[10444]: I1205 10:46:38.670818 10444 generic.go:334] "Generic (PLEG): container finished" podID="8d76404b-6d62-4a61-b6f6-0c8073eba198" containerID="163ef91a3ca09ff8a80da2c166f293609027eb6a5251a7136d985bf676587d9d" exitCode=137 Dec 05 10:46:38.671558 master-0 kubenswrapper[10444]: I1205 10:46:38.670853 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" event={"ID":"8d76404b-6d62-4a61-b6f6-0c8073eba198","Type":"ContainerDied","Data":"163ef91a3ca09ff8a80da2c166f293609027eb6a5251a7136d985bf676587d9d"} Dec 05 10:46:39.310334 master-0 kubenswrapper[10444]: I1205 10:46:39.310270 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-7dfc5b745f-67rx7_8d76404b-6d62-4a61-b6f6-0c8073eba198/multus-admission-controller/0.log" Dec 05 10:46:39.310550 master-0 kubenswrapper[10444]: I1205 10:46:39.310373 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:46:39.460893 master-0 kubenswrapper[10444]: I1205 10:46:39.460816 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") pod \"8d76404b-6d62-4a61-b6f6-0c8073eba198\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " Dec 05 10:46:39.460893 master-0 kubenswrapper[10444]: I1205 10:46:39.460904 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kb5c\" (UniqueName: \"kubernetes.io/projected/8d76404b-6d62-4a61-b6f6-0c8073eba198-kube-api-access-7kb5c\") pod \"8d76404b-6d62-4a61-b6f6-0c8073eba198\" (UID: \"8d76404b-6d62-4a61-b6f6-0c8073eba198\") " Dec 05 10:46:39.465341 master-0 kubenswrapper[10444]: I1205 10:46:39.465227 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "8d76404b-6d62-4a61-b6f6-0c8073eba198" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:46:39.466001 master-0 kubenswrapper[10444]: I1205 10:46:39.465928 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d76404b-6d62-4a61-b6f6-0c8073eba198-kube-api-access-7kb5c" (OuterVolumeSpecName: "kube-api-access-7kb5c") pod "8d76404b-6d62-4a61-b6f6-0c8073eba198" (UID: "8d76404b-6d62-4a61-b6f6-0c8073eba198"). InnerVolumeSpecName "kube-api-access-7kb5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:46:39.563793 master-0 kubenswrapper[10444]: I1205 10:46:39.563640 10444 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8d76404b-6d62-4a61-b6f6-0c8073eba198-webhook-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:39.563793 master-0 kubenswrapper[10444]: I1205 10:46:39.563711 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kb5c\" (UniqueName: \"kubernetes.io/projected/8d76404b-6d62-4a61-b6f6-0c8073eba198-kube-api-access-7kb5c\") on node \"master-0\" DevicePath \"\"" Dec 05 10:46:39.679303 master-0 kubenswrapper[10444]: I1205 10:46:39.679230 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-admission-controller-7dfc5b745f-67rx7_8d76404b-6d62-4a61-b6f6-0c8073eba198/multus-admission-controller/0.log" Dec 05 10:46:39.679852 master-0 kubenswrapper[10444]: I1205 10:46:39.679324 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" event={"ID":"8d76404b-6d62-4a61-b6f6-0c8073eba198","Type":"ContainerDied","Data":"e9475cd437d6e0adb87102a00093bc37d7716f625ab1f62c89982010f22e3b2c"} Dec 05 10:46:39.679852 master-0 kubenswrapper[10444]: I1205 10:46:39.679387 10444 scope.go:117] "RemoveContainer" containerID="d70a4de37cbf6d2709586eaecfa831636df1b7ccf32f015e3097c3c53b817cac" Dec 05 10:46:39.679852 master-0 kubenswrapper[10444]: I1205 10:46:39.679479 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-7dfc5b745f-67rx7" Dec 05 10:46:39.698622 master-0 kubenswrapper[10444]: I1205 10:46:39.698534 10444 scope.go:117] "RemoveContainer" containerID="163ef91a3ca09ff8a80da2c166f293609027eb6a5251a7136d985bf676587d9d" Dec 05 10:46:39.700989 master-0 kubenswrapper[10444]: I1205 10:46:39.700918 10444 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/multus-admission-controller-7dfc5b745f-67rx7"] Dec 05 10:46:39.704331 master-0 kubenswrapper[10444]: I1205 10:46:39.704251 10444 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/multus-admission-controller-7dfc5b745f-67rx7"] Dec 05 10:46:41.601670 master-0 kubenswrapper[10444]: I1205 10:46:41.601625 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d76404b-6d62-4a61-b6f6-0c8073eba198" path="/var/lib/kubelet/pods/8d76404b-6d62-4a61-b6f6-0c8073eba198/volumes" Dec 05 10:46:44.621680 master-0 kubenswrapper[10444]: I1205 10:46:44.621597 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/installer-2-master-0"] Dec 05 10:46:44.622950 master-0 kubenswrapper[10444]: E1205 10:46:44.621922 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d76404b-6d62-4a61-b6f6-0c8073eba198" containerName="multus-admission-controller" Dec 05 10:46:44.622950 master-0 kubenswrapper[10444]: I1205 10:46:44.621938 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d76404b-6d62-4a61-b6f6-0c8073eba198" containerName="multus-admission-controller" Dec 05 10:46:44.622950 master-0 kubenswrapper[10444]: E1205 10:46:44.621947 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1a50dc7-2abb-4183-a391-6f75ab05110c" containerName="installer" Dec 05 10:46:44.622950 master-0 kubenswrapper[10444]: I1205 10:46:44.621955 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1a50dc7-2abb-4183-a391-6f75ab05110c" containerName="installer" Dec 05 10:46:44.622950 master-0 kubenswrapper[10444]: E1205 10:46:44.621967 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d76404b-6d62-4a61-b6f6-0c8073eba198" containerName="kube-rbac-proxy" Dec 05 10:46:44.622950 master-0 kubenswrapper[10444]: I1205 10:46:44.621974 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d76404b-6d62-4a61-b6f6-0c8073eba198" containerName="kube-rbac-proxy" Dec 05 10:46:44.622950 master-0 kubenswrapper[10444]: I1205 10:46:44.622114 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d76404b-6d62-4a61-b6f6-0c8073eba198" containerName="kube-rbac-proxy" Dec 05 10:46:44.622950 master-0 kubenswrapper[10444]: I1205 10:46:44.622129 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1a50dc7-2abb-4183-a391-6f75ab05110c" containerName="installer" Dec 05 10:46:44.622950 master-0 kubenswrapper[10444]: I1205 10:46:44.622153 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d76404b-6d62-4a61-b6f6-0c8073eba198" containerName="multus-admission-controller" Dec 05 10:46:44.622950 master-0 kubenswrapper[10444]: I1205 10:46:44.622708 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-master-0" Dec 05 10:46:44.624781 master-0 kubenswrapper[10444]: I1205 10:46:44.624651 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd"/"kube-root-ca.crt" Dec 05 10:46:44.624781 master-0 kubenswrapper[10444]: I1205 10:46:44.624691 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd"/"installer-sa-dockercfg-76wfq" Dec 05 10:46:44.642106 master-0 kubenswrapper[10444]: I1205 10:46:44.641980 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd/installer-2-master-0"] Dec 05 10:46:44.733766 master-0 kubenswrapper[10444]: I1205 10:46:44.733680 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:46:44.733990 master-0 kubenswrapper[10444]: I1205 10:46:44.733793 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:46:44.734113 master-0 kubenswrapper[10444]: I1205 10:46:44.734056 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:46:44.835792 master-0 kubenswrapper[10444]: I1205 10:46:44.835735 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:46:44.835971 master-0 kubenswrapper[10444]: I1205 10:46:44.835807 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:46:44.835971 master-0 kubenswrapper[10444]: I1205 10:46:44.835886 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:46:44.836248 master-0 kubenswrapper[10444]: I1205 10:46:44.836149 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:46:44.836313 master-0 kubenswrapper[10444]: I1205 10:46:44.836252 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:46:44.851963 master-0 kubenswrapper[10444]: I1205 10:46:44.851890 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:46:44.953999 master-0 kubenswrapper[10444]: I1205 10:46:44.953809 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-master-0" Dec 05 10:46:45.386723 master-0 kubenswrapper[10444]: I1205 10:46:45.386663 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd/installer-2-master-0"] Dec 05 10:46:45.730741 master-0 kubenswrapper[10444]: I1205 10:46:45.730647 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-2-master-0" event={"ID":"68058d25-5888-495f-8571-453f0bf919d9","Type":"ContainerStarted","Data":"7533cfa355ed03950ce16c744c7022b947100926d9199c99ab99b289da6cb649"} Dec 05 10:46:45.959730 master-0 kubenswrapper[10444]: I1205 10:46:45.959622 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:46:46.740658 master-0 kubenswrapper[10444]: I1205 10:46:46.740579 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-2-master-0" event={"ID":"68058d25-5888-495f-8571-453f0bf919d9","Type":"ContainerStarted","Data":"2ea94dd01104138f23272416f98911388c3b177dca6ad44593587087deb08528"} Dec 05 10:46:46.869680 master-0 kubenswrapper[10444]: I1205 10:46:46.869548 10444 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/installer-2-master-0" podStartSLOduration=2.8695132340000002 podStartE2EDuration="2.869513234s" podCreationTimestamp="2025-12-05 10:46:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:46:46.867408492 +0000 UTC m=+547.459321109" watchObservedRunningTime="2025-12-05 10:46:46.869513234 +0000 UTC m=+547.461425831" Dec 05 10:47:10.917366 master-0 kubenswrapper[10444]: I1205 10:47:10.917306 10444 generic.go:334] "Generic (PLEG): container finished" podID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerID="df3d2e1cdc52294902ada6e0e872265f937f9cd41ea5e7643f484cd5a4774ce3" exitCode=0 Dec 05 10:47:10.917366 master-0 kubenswrapper[10444]: I1205 10:47:10.917353 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" event={"ID":"445d75af-d072-4fa0-91a7-f3fa579b9ca9","Type":"ContainerDied","Data":"df3d2e1cdc52294902ada6e0e872265f937f9cd41ea5e7643f484cd5a4774ce3"} Dec 05 10:47:10.918032 master-0 kubenswrapper[10444]: I1205 10:47:10.917382 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" event={"ID":"445d75af-d072-4fa0-91a7-f3fa579b9ca9","Type":"ContainerStarted","Data":"4d0f230cfcdcef315da004bcdd2084b9e620df4b288a3ee54899132f694ed0c0"} Dec 05 10:47:10.918032 master-0 kubenswrapper[10444]: I1205 10:47:10.917398 10444 scope.go:117] "RemoveContainer" containerID="dfc1c20819ce4f467461d89e6abf68fcdb1202febbeb18490a6b3dcd2aec085f" Dec 05 10:47:11.556516 master-0 kubenswrapper[10444]: I1205 10:47:11.556378 10444 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:11.559362 master-0 kubenswrapper[10444]: I1205 10:47:11.559302 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:47:11.559362 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:47:11.559362 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:47:11.559362 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:47:11.559688 master-0 kubenswrapper[10444]: I1205 10:47:11.559378 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:47:12.406505 master-0 kubenswrapper[10444]: I1205 10:47:12.404219 10444 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/telemeter-client-86cb595668-52qnw"] Dec 05 10:47:12.406505 master-0 kubenswrapper[10444]: I1205 10:47:12.406394 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.411194 master-0 kubenswrapper[10444]: I1205 10:47:12.411148 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client" Dec 05 10:47:12.411411 master-0 kubenswrapper[10444]: I1205 10:47:12.411391 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client-tls" Dec 05 10:47:12.411590 master-0 kubenswrapper[10444]: I1205 10:47:12.411561 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client-dockercfg-2kvvs" Dec 05 10:47:12.411871 master-0 kubenswrapper[10444]: I1205 10:47:12.411808 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemeter-client-serving-certs-ca-bundle" Dec 05 10:47:12.412052 master-0 kubenswrapper[10444]: I1205 10:47:12.412013 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"federate-client-certs" Dec 05 10:47:12.412187 master-0 kubenswrapper[10444]: I1205 10:47:12.412145 10444 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client-kube-rbac-proxy-config" Dec 05 10:47:12.420489 master-0 kubenswrapper[10444]: I1205 10:47:12.420403 10444 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemeter-trusted-ca-bundle-56c9b9fa8d9gs" Dec 05 10:47:12.426119 master-0 kubenswrapper[10444]: I1205 10:47:12.426054 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/telemeter-client-86cb595668-52qnw"] Dec 05 10:47:12.550599 master-0 kubenswrapper[10444]: I1205 10:47:12.550521 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.550599 master-0 kubenswrapper[10444]: I1205 10:47:12.550582 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.550870 master-0 kubenswrapper[10444]: I1205 10:47:12.550633 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.550870 master-0 kubenswrapper[10444]: I1205 10:47:12.550656 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-metrics-client-ca\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.550963 master-0 kubenswrapper[10444]: I1205 10:47:12.550849 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.551105 master-0 kubenswrapper[10444]: I1205 10:47:12.551061 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgtnt\" (UniqueName: \"kubernetes.io/projected/384f1c47-f511-4b58-aa7f-71aef6ef91a9-kube-api-access-lgtnt\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.551158 master-0 kubenswrapper[10444]: I1205 10:47:12.551147 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.551213 master-0 kubenswrapper[10444]: I1205 10:47:12.551175 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.558161 master-0 kubenswrapper[10444]: I1205 10:47:12.558120 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:47:12.558161 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:47:12.558161 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:47:12.558161 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:47:12.558305 master-0 kubenswrapper[10444]: I1205 10:47:12.558176 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:47:12.652289 master-0 kubenswrapper[10444]: I1205 10:47:12.652211 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.652289 master-0 kubenswrapper[10444]: I1205 10:47:12.652279 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-metrics-client-ca\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.652651 master-0 kubenswrapper[10444]: I1205 10:47:12.652588 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.652774 master-0 kubenswrapper[10444]: I1205 10:47:12.652716 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgtnt\" (UniqueName: \"kubernetes.io/projected/384f1c47-f511-4b58-aa7f-71aef6ef91a9-kube-api-access-lgtnt\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.652820 master-0 kubenswrapper[10444]: I1205 10:47:12.652791 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.653079 master-0 kubenswrapper[10444]: I1205 10:47:12.652988 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.653314 master-0 kubenswrapper[10444]: I1205 10:47:12.653128 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.653314 master-0 kubenswrapper[10444]: I1205 10:47:12.653172 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.653417 master-0 kubenswrapper[10444]: I1205 10:47:12.653307 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-metrics-client-ca\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.654034 master-0 kubenswrapper[10444]: I1205 10:47:12.653983 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.654641 master-0 kubenswrapper[10444]: I1205 10:47:12.654603 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.655739 master-0 kubenswrapper[10444]: I1205 10:47:12.655680 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.656003 master-0 kubenswrapper[10444]: I1205 10:47:12.655965 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.656636 master-0 kubenswrapper[10444]: I1205 10:47:12.656537 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.657018 master-0 kubenswrapper[10444]: I1205 10:47:12.656962 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.674230 master-0 kubenswrapper[10444]: I1205 10:47:12.674165 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgtnt\" (UniqueName: \"kubernetes.io/projected/384f1c47-f511-4b58-aa7f-71aef6ef91a9-kube-api-access-lgtnt\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:12.726518 master-0 kubenswrapper[10444]: I1205 10:47:12.726460 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:13.162375 master-0 kubenswrapper[10444]: I1205 10:47:13.162302 10444 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/telemeter-client-86cb595668-52qnw"] Dec 05 10:47:13.167871 master-0 kubenswrapper[10444]: W1205 10:47:13.167787 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod384f1c47_f511_4b58_aa7f_71aef6ef91a9.slice/crio-7f2b0f7328f1fdf55c43aa55c01436ada2488903a3cf2d212fd3c7469222fb7b WatchSource:0}: Error finding container 7f2b0f7328f1fdf55c43aa55c01436ada2488903a3cf2d212fd3c7469222fb7b: Status 404 returned error can't find the container with id 7f2b0f7328f1fdf55c43aa55c01436ada2488903a3cf2d212fd3c7469222fb7b Dec 05 10:47:13.171000 master-0 kubenswrapper[10444]: I1205 10:47:13.170951 10444 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:47:13.558230 master-0 kubenswrapper[10444]: I1205 10:47:13.558155 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:47:13.558230 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:47:13.558230 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:47:13.558230 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:47:13.558932 master-0 kubenswrapper[10444]: I1205 10:47:13.558236 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:47:13.956769 master-0 kubenswrapper[10444]: I1205 10:47:13.956485 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" event={"ID":"384f1c47-f511-4b58-aa7f-71aef6ef91a9","Type":"ContainerStarted","Data":"7f2b0f7328f1fdf55c43aa55c01436ada2488903a3cf2d212fd3c7469222fb7b"} Dec 05 10:47:14.557973 master-0 kubenswrapper[10444]: I1205 10:47:14.557914 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:47:14.557973 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:47:14.557973 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:47:14.557973 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:47:14.558865 master-0 kubenswrapper[10444]: I1205 10:47:14.557977 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:47:14.964448 master-0 kubenswrapper[10444]: I1205 10:47:14.964296 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/3.log" Dec 05 10:47:14.965287 master-0 kubenswrapper[10444]: I1205 10:47:14.965215 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/2.log" Dec 05 10:47:14.965634 master-0 kubenswrapper[10444]: I1205 10:47:14.965569 10444 generic.go:334] "Generic (PLEG): container finished" podID="22676fac-b770-4937-9bee-7478bd1babb7" containerID="76b36235b0c1dc04c6934f55253404fa8e8aa924bd0ff655fd03d44b54d22caa" exitCode=1 Dec 05 10:47:14.965634 master-0 kubenswrapper[10444]: I1205 10:47:14.965603 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" event={"ID":"22676fac-b770-4937-9bee-7478bd1babb7","Type":"ContainerDied","Data":"76b36235b0c1dc04c6934f55253404fa8e8aa924bd0ff655fd03d44b54d22caa"} Dec 05 10:47:14.965634 master-0 kubenswrapper[10444]: I1205 10:47:14.965635 10444 scope.go:117] "RemoveContainer" containerID="0eefd6344cec169209984eb638655b84736a398edd25e1b16bc1da6cc4d7c7d9" Dec 05 10:47:14.966671 master-0 kubenswrapper[10444]: I1205 10:47:14.966356 10444 scope.go:117] "RemoveContainer" containerID="76b36235b0c1dc04c6934f55253404fa8e8aa924bd0ff655fd03d44b54d22caa" Dec 05 10:47:14.967159 master-0 kubenswrapper[10444]: E1205 10:47:14.966714 10444 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ingress-operator\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ingress-operator pod=ingress-operator-8649c48786-cgt5x_openshift-ingress-operator(22676fac-b770-4937-9bee-7478bd1babb7)\"" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" podUID="22676fac-b770-4937-9bee-7478bd1babb7" Dec 05 10:47:14.996548 master-0 kubenswrapper[10444]: I1205 10:47:14.995137 10444 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0"] Dec 05 10:47:14.997483 master-0 kubenswrapper[10444]: I1205 10:47:14.997415 10444 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/bootstrap-kube-apiserver-master-0"] Dec 05 10:47:14.997561 master-0 kubenswrapper[10444]: I1205 10:47:14.997485 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:14.997760 master-0 kubenswrapper[10444]: I1205 10:47:14.997654 10444 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-master-0"] Dec 05 10:47:14.997760 master-0 kubenswrapper[10444]: I1205 10:47:14.997711 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" podUID="d75143d9bc4a2dc15781dc51ccff632a" containerName="kube-apiserver" containerID="cri-o://10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298" gracePeriod=15 Dec 05 10:47:14.997918 master-0 kubenswrapper[10444]: I1205 10:47:14.997885 10444 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" podUID="d75143d9bc4a2dc15781dc51ccff632a" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef" gracePeriod=15 Dec 05 10:47:14.998028 master-0 kubenswrapper[10444]: E1205 10:47:14.998008 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d75143d9bc4a2dc15781dc51ccff632a" containerName="kube-apiserver-insecure-readyz" Dec 05 10:47:14.998075 master-0 kubenswrapper[10444]: I1205 10:47:14.998030 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="d75143d9bc4a2dc15781dc51ccff632a" containerName="kube-apiserver-insecure-readyz" Dec 05 10:47:14.998075 master-0 kubenswrapper[10444]: E1205 10:47:14.998043 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d75143d9bc4a2dc15781dc51ccff632a" containerName="setup" Dec 05 10:47:14.998075 master-0 kubenswrapper[10444]: I1205 10:47:14.998052 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="d75143d9bc4a2dc15781dc51ccff632a" containerName="setup" Dec 05 10:47:14.998075 master-0 kubenswrapper[10444]: E1205 10:47:14.998072 10444 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d75143d9bc4a2dc15781dc51ccff632a" containerName="kube-apiserver" Dec 05 10:47:14.998239 master-0 kubenswrapper[10444]: I1205 10:47:14.998080 10444 state_mem.go:107] "Deleted CPUSet assignment" podUID="d75143d9bc4a2dc15781dc51ccff632a" containerName="kube-apiserver" Dec 05 10:47:14.998239 master-0 kubenswrapper[10444]: I1205 10:47:14.998209 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="d75143d9bc4a2dc15781dc51ccff632a" containerName="setup" Dec 05 10:47:14.998239 master-0 kubenswrapper[10444]: I1205 10:47:14.998226 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="d75143d9bc4a2dc15781dc51ccff632a" containerName="kube-apiserver-insecure-readyz" Dec 05 10:47:14.998355 master-0 kubenswrapper[10444]: I1205 10:47:14.998246 10444 memory_manager.go:354] "RemoveStaleState removing state" podUID="d75143d9bc4a2dc15781dc51ccff632a" containerName="kube-apiserver" Dec 05 10:47:15.017633 master-0 kubenswrapper[10444]: I1205 10:47:15.017557 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.049230 master-0 kubenswrapper[10444]: E1205 10:47:15.047726 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 192.168.32.10:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.056635 master-0 kubenswrapper[10444]: E1205 10:47:15.056566 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 192.168.32.10:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.086966 master-0 kubenswrapper[10444]: I1205 10:47:15.086913 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-log\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.087056 master-0 kubenswrapper[10444]: I1205 10:47:15.086976 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-audit-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.087056 master-0 kubenswrapper[10444]: I1205 10:47:15.087018 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.087056 master-0 kubenswrapper[10444]: I1205 10:47:15.087038 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-manifests\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.087146 master-0 kubenswrapper[10444]: I1205 10:47:15.087059 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-cert-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.087146 master-0 kubenswrapper[10444]: I1205 10:47:15.087093 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-lock\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.087416 master-0 kubenswrapper[10444]: I1205 10:47:15.087363 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-resource-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.087466 master-0 kubenswrapper[10444]: I1205 10:47:15.087445 10444 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.113347 master-0 kubenswrapper[10444]: E1205 10:47:15.113105 10444 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-monitoring/events\": dial tcp 192.168.32.10:6443: connect: connection refused" event="&Event{ObjectMeta:{telemeter-client-86cb595668-52qnw.187e4bf89ba72c87 openshift-monitoring 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-monitoring,Name:telemeter-client-86cb595668-52qnw,UID:384f1c47-f511-4b58-aa7f-71aef6ef91a9,APIVersion:v1,ResourceVersion:11850,FieldPath:spec.containers{telemeter-client},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:445efcbc0255b904e1584fe9be9a513c1a9784088e35dd0abbdff5cae0961861\" in 1.94s (1.94s including waiting). Image size: 474996496 bytes.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:47:15.111210119 +0000 UTC m=+575.703122686,LastTimestamp:2025-12-05 10:47:15.111210119 +0000 UTC m=+575.703122686,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:47:15.188224 master-0 kubenswrapper[10444]: I1205 10:47:15.188179 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-resource-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.188483 master-0 kubenswrapper[10444]: I1205 10:47:15.188461 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.188620 master-0 kubenswrapper[10444]: I1205 10:47:15.188294 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-resource-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.188705 master-0 kubenswrapper[10444]: I1205 10:47:15.188601 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-log\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.188835 master-0 kubenswrapper[10444]: I1205 10:47:15.188816 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-audit-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.188987 master-0 kubenswrapper[10444]: I1205 10:47:15.188940 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.189106 master-0 kubenswrapper[10444]: I1205 10:47:15.189089 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-manifests\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.189229 master-0 kubenswrapper[10444]: I1205 10:47:15.189212 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-cert-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.189365 master-0 kubenswrapper[10444]: I1205 10:47:15.189345 10444 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-lock\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.189529 master-0 kubenswrapper[10444]: I1205 10:47:15.189002 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.189652 master-0 kubenswrapper[10444]: I1205 10:47:15.188700 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.189740 master-0 kubenswrapper[10444]: I1205 10:47:15.189160 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-manifests\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.189833 master-0 kubenswrapper[10444]: I1205 10:47:15.188862 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-audit-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.189930 master-0 kubenswrapper[10444]: I1205 10:47:15.189266 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-cert-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.190057 master-0 kubenswrapper[10444]: I1205 10:47:15.188715 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-log\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.190150 master-0 kubenswrapper[10444]: I1205 10:47:15.189386 10444 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-lock\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.348871 master-0 kubenswrapper[10444]: I1205 10:47:15.348802 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.357700 master-0 kubenswrapper[10444]: I1205 10:47:15.357641 10444 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.385449 master-0 kubenswrapper[10444]: W1205 10:47:15.385278 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda906debd0c35952850935aee2d607cce.slice/crio-4446e01522dc26b79f498f8e0dc137457645ab081cac6529b44e9b557b9d72f8 WatchSource:0}: Error finding container 4446e01522dc26b79f498f8e0dc137457645ab081cac6529b44e9b557b9d72f8: Status 404 returned error can't find the container with id 4446e01522dc26b79f498f8e0dc137457645ab081cac6529b44e9b557b9d72f8 Dec 05 10:47:15.386947 master-0 kubenswrapper[10444]: W1205 10:47:15.386895 10444 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb89698aa356a3bc32694e2b098f9a900.slice/crio-b94fba3ebe8eacc577514add51b89ec37946322bc96bbe0ee6d3e8d5b7830ae3 WatchSource:0}: Error finding container b94fba3ebe8eacc577514add51b89ec37946322bc96bbe0ee6d3e8d5b7830ae3: Status 404 returned error can't find the container with id b94fba3ebe8eacc577514add51b89ec37946322bc96bbe0ee6d3e8d5b7830ae3 Dec 05 10:47:15.559964 master-0 kubenswrapper[10444]: I1205 10:47:15.559843 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:47:15.559964 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:47:15.559964 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:47:15.559964 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:47:15.559964 master-0 kubenswrapper[10444]: I1205 10:47:15.559915 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:47:15.976874 master-0 kubenswrapper[10444]: I1205 10:47:15.976786 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" event={"ID":"384f1c47-f511-4b58-aa7f-71aef6ef91a9","Type":"ContainerStarted","Data":"8efd6b5426f7f4e0c4afaad057a5579d54ff332cc348b626c259c2a524e918a9"} Dec 05 10:47:15.980491 master-0 kubenswrapper[10444]: I1205 10:47:15.980443 10444 generic.go:334] "Generic (PLEG): container finished" podID="b89698aa356a3bc32694e2b098f9a900" containerID="c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b" exitCode=0 Dec 05 10:47:15.980619 master-0 kubenswrapper[10444]: I1205 10:47:15.980464 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerDied","Data":"c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b"} Dec 05 10:47:15.980619 master-0 kubenswrapper[10444]: I1205 10:47:15.980530 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"b94fba3ebe8eacc577514add51b89ec37946322bc96bbe0ee6d3e8d5b7830ae3"} Dec 05 10:47:15.981765 master-0 kubenswrapper[10444]: E1205 10:47:15.981709 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 192.168.32.10:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:15.984303 master-0 kubenswrapper[10444]: I1205 10:47:15.984275 10444 generic.go:334] "Generic (PLEG): container finished" podID="d75143d9bc4a2dc15781dc51ccff632a" containerID="2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef" exitCode=0 Dec 05 10:47:15.986498 master-0 kubenswrapper[10444]: I1205 10:47:15.986462 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" event={"ID":"a906debd0c35952850935aee2d607cce","Type":"ContainerStarted","Data":"8930f7af011c1e43a64ba6a4ca85659538a4774816c777a7c34e7a4b10c3cab3"} Dec 05 10:47:15.986630 master-0 kubenswrapper[10444]: I1205 10:47:15.986513 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" event={"ID":"a906debd0c35952850935aee2d607cce","Type":"ContainerStarted","Data":"4446e01522dc26b79f498f8e0dc137457645ab081cac6529b44e9b557b9d72f8"} Dec 05 10:47:15.987749 master-0 kubenswrapper[10444]: E1205 10:47:15.987704 10444 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 192.168.32.10:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:15.988233 master-0 kubenswrapper[10444]: I1205 10:47:15.988200 10444 generic.go:334] "Generic (PLEG): container finished" podID="75ba20a1-b322-4ce8-88d2-d8ee210cc27a" containerID="c8f54c8d7593dfbeebbefc04c820ade3fd814ef907ea5e95f2df34b8ed43589a" exitCode=0 Dec 05 10:47:15.988315 master-0 kubenswrapper[10444]: I1205 10:47:15.988266 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-3-master-0" event={"ID":"75ba20a1-b322-4ce8-88d2-d8ee210cc27a","Type":"ContainerDied","Data":"c8f54c8d7593dfbeebbefc04c820ade3fd814ef907ea5e95f2df34b8ed43589a"} Dec 05 10:47:15.989492 master-0 kubenswrapper[10444]: I1205 10:47:15.989392 10444 status_manager.go:851] "Failed to get status for pod" podUID="75ba20a1-b322-4ce8-88d2-d8ee210cc27a" pod="openshift-kube-apiserver/installer-3-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-3-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:47:15.991574 master-0 kubenswrapper[10444]: I1205 10:47:15.991543 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/3.log" Dec 05 10:47:16.561140 master-0 kubenswrapper[10444]: I1205 10:47:16.560925 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:47:16.561140 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:47:16.561140 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:47:16.561140 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:47:16.561140 master-0 kubenswrapper[10444]: I1205 10:47:16.561006 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:47:17.008121 master-0 kubenswrapper[10444]: I1205 10:47:17.008056 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56"} Dec 05 10:47:17.012388 master-0 kubenswrapper[10444]: I1205 10:47:17.012355 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-2-master-0_68058d25-5888-495f-8571-453f0bf919d9/installer/0.log" Dec 05 10:47:17.012509 master-0 kubenswrapper[10444]: I1205 10:47:17.012459 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-2-master-0" event={"ID":"68058d25-5888-495f-8571-453f0bf919d9","Type":"ContainerDied","Data":"2ea94dd01104138f23272416f98911388c3b177dca6ad44593587087deb08528"} Dec 05 10:47:17.013015 master-0 kubenswrapper[10444]: I1205 10:47:17.012984 10444 generic.go:334] "Generic (PLEG): container finished" podID="68058d25-5888-495f-8571-453f0bf919d9" containerID="2ea94dd01104138f23272416f98911388c3b177dca6ad44593587087deb08528" exitCode=1 Dec 05 10:47:17.387857 master-0 kubenswrapper[10444]: I1205 10:47:17.387771 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:17.524523 master-0 kubenswrapper[10444]: I1205 10:47:17.524468 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " Dec 05 10:47:17.524641 master-0 kubenswrapper[10444]: I1205 10:47:17.524560 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir\") pod \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " Dec 05 10:47:17.524641 master-0 kubenswrapper[10444]: I1205 10:47:17.524611 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock\") pod \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " Dec 05 10:47:17.524915 master-0 kubenswrapper[10444]: I1205 10:47:17.524865 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "75ba20a1-b322-4ce8-88d2-d8ee210cc27a" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:17.525016 master-0 kubenswrapper[10444]: I1205 10:47:17.524968 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock" (OuterVolumeSpecName: "var-lock") pod "75ba20a1-b322-4ce8-88d2-d8ee210cc27a" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:17.528367 master-0 kubenswrapper[10444]: I1205 10:47:17.527564 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "75ba20a1-b322-4ce8-88d2-d8ee210cc27a" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:47:17.559900 master-0 kubenswrapper[10444]: I1205 10:47:17.559854 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:47:17.559900 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:47:17.559900 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:47:17.559900 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:47:17.560148 master-0 kubenswrapper[10444]: I1205 10:47:17.559925 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:47:17.627829 master-0 kubenswrapper[10444]: I1205 10:47:17.627521 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:17.627829 master-0 kubenswrapper[10444]: I1205 10:47:17.627750 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:17.627829 master-0 kubenswrapper[10444]: I1205 10:47:17.627760 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:17.979284 master-0 kubenswrapper[10444]: I1205 10:47:17.979151 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:47:18.095077 master-0 kubenswrapper[10444]: I1205 10:47:18.095019 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-audit-dir\") pod \"d75143d9bc4a2dc15781dc51ccff632a\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " Dec 05 10:47:18.095077 master-0 kubenswrapper[10444]: I1205 10:47:18.095079 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-etc-kubernetes-cloud\") pod \"d75143d9bc4a2dc15781dc51ccff632a\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " Dec 05 10:47:18.095324 master-0 kubenswrapper[10444]: I1205 10:47:18.095129 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-ssl-certs-host\") pod \"d75143d9bc4a2dc15781dc51ccff632a\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " Dec 05 10:47:18.095324 master-0 kubenswrapper[10444]: I1205 10:47:18.095152 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-secrets\") pod \"d75143d9bc4a2dc15781dc51ccff632a\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " Dec 05 10:47:18.095324 master-0 kubenswrapper[10444]: I1205 10:47:18.095197 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-config\") pod \"d75143d9bc4a2dc15781dc51ccff632a\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " Dec 05 10:47:18.095324 master-0 kubenswrapper[10444]: I1205 10:47:18.095221 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-logs\") pod \"d75143d9bc4a2dc15781dc51ccff632a\" (UID: \"d75143d9bc4a2dc15781dc51ccff632a\") " Dec 05 10:47:18.100673 master-0 kubenswrapper[10444]: I1205 10:47:18.100543 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-etc-kubernetes-cloud" (OuterVolumeSpecName: "etc-kubernetes-cloud") pod "d75143d9bc4a2dc15781dc51ccff632a" (UID: "d75143d9bc4a2dc15781dc51ccff632a"). InnerVolumeSpecName "etc-kubernetes-cloud". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:18.101490 master-0 kubenswrapper[10444]: I1205 10:47:18.101459 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "d75143d9bc4a2dc15781dc51ccff632a" (UID: "d75143d9bc4a2dc15781dc51ccff632a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:18.101692 master-0 kubenswrapper[10444]: I1205 10:47:18.101636 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-secrets" (OuterVolumeSpecName: "secrets") pod "d75143d9bc4a2dc15781dc51ccff632a" (UID: "d75143d9bc4a2dc15781dc51ccff632a"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:18.101789 master-0 kubenswrapper[10444]: I1205 10:47:18.101695 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-ssl-certs-host" (OuterVolumeSpecName: "ssl-certs-host") pod "d75143d9bc4a2dc15781dc51ccff632a" (UID: "d75143d9bc4a2dc15781dc51ccff632a"). InnerVolumeSpecName "ssl-certs-host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:18.113020 master-0 kubenswrapper[10444]: I1205 10:47:18.103978 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-config" (OuterVolumeSpecName: "config") pod "d75143d9bc4a2dc15781dc51ccff632a" (UID: "d75143d9bc4a2dc15781dc51ccff632a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:18.113020 master-0 kubenswrapper[10444]: I1205 10:47:18.108017 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-logs" (OuterVolumeSpecName: "logs") pod "d75143d9bc4a2dc15781dc51ccff632a" (UID: "d75143d9bc4a2dc15781dc51ccff632a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:18.113020 master-0 kubenswrapper[10444]: I1205 10:47:18.111682 10444 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-secrets\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:18.113020 master-0 kubenswrapper[10444]: I1205 10:47:18.111718 10444 reconciler_common.go:293] "Volume detached for volume \"ssl-certs-host\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-ssl-certs-host\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:18.113020 master-0 kubenswrapper[10444]: I1205 10:47:18.111733 10444 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:18.113020 master-0 kubenswrapper[10444]: I1205 10:47:18.111746 10444 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:18.113020 master-0 kubenswrapper[10444]: I1205 10:47:18.111758 10444 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-audit-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:18.113020 master-0 kubenswrapper[10444]: I1205 10:47:18.111770 10444 reconciler_common.go:293] "Volume detached for volume \"etc-kubernetes-cloud\" (UniqueName: \"kubernetes.io/host-path/d75143d9bc4a2dc15781dc51ccff632a-etc-kubernetes-cloud\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:18.172233 master-0 kubenswrapper[10444]: I1205 10:47:18.172017 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5"} Dec 05 10:47:18.172233 master-0 kubenswrapper[10444]: I1205 10:47:18.172084 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf"} Dec 05 10:47:18.172233 master-0 kubenswrapper[10444]: I1205 10:47:18.172100 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63"} Dec 05 10:47:18.179368 master-0 kubenswrapper[10444]: I1205 10:47:18.179320 10444 generic.go:334] "Generic (PLEG): container finished" podID="d75143d9bc4a2dc15781dc51ccff632a" containerID="10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298" exitCode=0 Dec 05 10:47:18.179557 master-0 kubenswrapper[10444]: I1205 10:47:18.179397 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" Dec 05 10:47:18.179725 master-0 kubenswrapper[10444]: I1205 10:47:18.179694 10444 scope.go:117] "RemoveContainer" containerID="2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef" Dec 05 10:47:18.182091 master-0 kubenswrapper[10444]: I1205 10:47:18.182055 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-3-master-0" event={"ID":"75ba20a1-b322-4ce8-88d2-d8ee210cc27a","Type":"ContainerDied","Data":"3b1988a80333a5d160a32b470d3b1aa05cd5259d79366c745b970eff4d6866a0"} Dec 05 10:47:18.182091 master-0 kubenswrapper[10444]: I1205 10:47:18.182078 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:18.182091 master-0 kubenswrapper[10444]: I1205 10:47:18.182092 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b1988a80333a5d160a32b470d3b1aa05cd5259d79366c745b970eff4d6866a0" Dec 05 10:47:18.185479 master-0 kubenswrapper[10444]: I1205 10:47:18.185412 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" event={"ID":"384f1c47-f511-4b58-aa7f-71aef6ef91a9","Type":"ContainerStarted","Data":"e612b19ac4e57f925111fab12ca147d4c2a56896d21a8e251987dcff936584f2"} Dec 05 10:47:18.185479 master-0 kubenswrapper[10444]: I1205 10:47:18.185455 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" event={"ID":"384f1c47-f511-4b58-aa7f-71aef6ef91a9","Type":"ContainerStarted","Data":"1c56615662598432e2d070e8fe4be435799899ad38d22673ccf16e74d23717a9"} Dec 05 10:47:18.221460 master-0 kubenswrapper[10444]: I1205 10:47:18.221293 10444 scope.go:117] "RemoveContainer" containerID="10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298" Dec 05 10:47:18.238474 master-0 kubenswrapper[10444]: I1205 10:47:18.238439 10444 scope.go:117] "RemoveContainer" containerID="738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948" Dec 05 10:47:18.286477 master-0 kubenswrapper[10444]: I1205 10:47:18.286447 10444 scope.go:117] "RemoveContainer" containerID="2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef" Dec 05 10:47:18.291082 master-0 kubenswrapper[10444]: E1205 10:47:18.291043 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef\": container with ID starting with 2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef not found: ID does not exist" containerID="2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef" Dec 05 10:47:18.291465 master-0 kubenswrapper[10444]: I1205 10:47:18.291278 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef"} err="failed to get container status \"2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef\": rpc error: code = NotFound desc = could not find container \"2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef\": container with ID starting with 2775fe6de66af264585de56a28a806f17a841a0b070c2c61c7ba00db6c1bf6ef not found: ID does not exist" Dec 05 10:47:18.291562 master-0 kubenswrapper[10444]: I1205 10:47:18.291549 10444 scope.go:117] "RemoveContainer" containerID="10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298" Dec 05 10:47:18.294848 master-0 kubenswrapper[10444]: E1205 10:47:18.294815 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298\": container with ID starting with 10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298 not found: ID does not exist" containerID="10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298" Dec 05 10:47:18.295082 master-0 kubenswrapper[10444]: I1205 10:47:18.294980 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298"} err="failed to get container status \"10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298\": rpc error: code = NotFound desc = could not find container \"10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298\": container with ID starting with 10f08ef8e834a8e2905c69479a0f8ddbdb4baa94cf51bae8788216ba6f1c8298 not found: ID does not exist" Dec 05 10:47:18.295199 master-0 kubenswrapper[10444]: I1205 10:47:18.295187 10444 scope.go:117] "RemoveContainer" containerID="738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948" Dec 05 10:47:18.295759 master-0 kubenswrapper[10444]: E1205 10:47:18.295718 10444 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948\": container with ID starting with 738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948 not found: ID does not exist" containerID="738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948" Dec 05 10:47:18.295759 master-0 kubenswrapper[10444]: I1205 10:47:18.295742 10444 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948"} err="failed to get container status \"738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948\": rpc error: code = NotFound desc = could not find container \"738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948\": container with ID starting with 738bb4c18cb4e63eb85f83e734199cff6ee909754cc0c041a2d2e2c296b9e948 not found: ID does not exist" Dec 05 10:47:18.533969 master-0 kubenswrapper[10444]: I1205 10:47:18.533916 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-2-master-0_68058d25-5888-495f-8571-453f0bf919d9/installer/0.log" Dec 05 10:47:18.533969 master-0 kubenswrapper[10444]: I1205 10:47:18.533980 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:18.567340 master-0 kubenswrapper[10444]: I1205 10:47:18.567037 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:47:18.567340 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:47:18.567340 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:47:18.567340 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:47:18.567340 master-0 kubenswrapper[10444]: I1205 10:47:18.567095 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:47:18.624471 master-0 kubenswrapper[10444]: I1205 10:47:18.623955 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir\") pod \"68058d25-5888-495f-8571-453f0bf919d9\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " Dec 05 10:47:18.624471 master-0 kubenswrapper[10444]: I1205 10:47:18.624108 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock\") pod \"68058d25-5888-495f-8571-453f0bf919d9\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " Dec 05 10:47:18.624471 master-0 kubenswrapper[10444]: I1205 10:47:18.624097 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "68058d25-5888-495f-8571-453f0bf919d9" (UID: "68058d25-5888-495f-8571-453f0bf919d9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:18.624471 master-0 kubenswrapper[10444]: I1205 10:47:18.624254 10444 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"68058d25-5888-495f-8571-453f0bf919d9\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " Dec 05 10:47:18.624471 master-0 kubenswrapper[10444]: I1205 10:47:18.624128 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock" (OuterVolumeSpecName: "var-lock") pod "68058d25-5888-495f-8571-453f0bf919d9" (UID: "68058d25-5888-495f-8571-453f0bf919d9"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:18.625019 master-0 kubenswrapper[10444]: I1205 10:47:18.624786 10444 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:18.625019 master-0 kubenswrapper[10444]: I1205 10:47:18.624805 10444 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:18.628479 master-0 kubenswrapper[10444]: I1205 10:47:18.627690 10444 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "68058d25-5888-495f-8571-453f0bf919d9" (UID: "68058d25-5888-495f-8571-453f0bf919d9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:47:18.726974 master-0 kubenswrapper[10444]: I1205 10:47:18.726887 10444 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:19.195195 master-0 kubenswrapper[10444]: I1205 10:47:19.195100 10444 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-2-master-0_68058d25-5888-495f-8571-453f0bf919d9/installer/0.log" Dec 05 10:47:19.195195 master-0 kubenswrapper[10444]: I1205 10:47:19.195186 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-2-master-0" event={"ID":"68058d25-5888-495f-8571-453f0bf919d9","Type":"ContainerDied","Data":"7533cfa355ed03950ce16c744c7022b947100926d9199c99ab99b289da6cb649"} Dec 05 10:47:19.195195 master-0 kubenswrapper[10444]: I1205 10:47:19.195210 10444 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:19.196027 master-0 kubenswrapper[10444]: I1205 10:47:19.195213 10444 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7533cfa355ed03950ce16c744c7022b947100926d9199c99ab99b289da6cb649" Dec 05 10:47:19.200203 master-0 kubenswrapper[10444]: I1205 10:47:19.200150 10444 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5"} Dec 05 10:47:19.200436 master-0 kubenswrapper[10444]: I1205 10:47:19.200375 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:19.557644 master-0 kubenswrapper[10444]: I1205 10:47:19.557539 10444 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:19.561527 master-0 kubenswrapper[10444]: I1205 10:47:19.561339 10444 patch_prober.go:28] interesting pod/router-default-5465c8b4db-s4c2f container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 10:47:19.561527 master-0 kubenswrapper[10444]: [-]has-synced failed: reason withheld Dec 05 10:47:19.561527 master-0 kubenswrapper[10444]: [+]process-running ok Dec 05 10:47:19.561527 master-0 kubenswrapper[10444]: healthz check failed Dec 05 10:47:19.561527 master-0 kubenswrapper[10444]: I1205 10:47:19.561450 10444 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" podUID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:47:19.607761 master-0 kubenswrapper[10444]: I1205 10:47:19.607689 10444 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d75143d9bc4a2dc15781dc51ccff632a" path="/var/lib/kubelet/pods/d75143d9bc4a2dc15781dc51ccff632a/volumes" Dec 05 10:47:19.608144 master-0 kubenswrapper[10444]: I1205 10:47:19.608104 10444 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/bootstrap-kube-apiserver-master-0" podUID="" Dec 05 10:47:19.911815 master-0 systemd[1]: Stopping Kubernetes Kubelet... Dec 05 10:47:19.928462 master-0 systemd[1]: kubelet.service: Deactivated successfully. Dec 05 10:47:19.928717 master-0 systemd[1]: Stopped Kubernetes Kubelet. Dec 05 10:47:19.929740 master-0 systemd[1]: kubelet.service: Consumed 1min 20.735s CPU time. Dec 05 10:47:19.952956 master-0 systemd[1]: Starting Kubernetes Kubelet... Dec 05 10:47:20.029726 master-0 kubenswrapper[24928]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:47:20.029726 master-0 kubenswrapper[24928]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 05 10:47:20.029726 master-0 kubenswrapper[24928]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:47:20.029726 master-0 kubenswrapper[24928]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:47:20.029726 master-0 kubenswrapper[24928]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 05 10:47:20.029726 master-0 kubenswrapper[24928]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 10:47:20.030364 master-0 kubenswrapper[24928]: I1205 10:47:20.029819 24928 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 05 10:47:20.031931 master-0 kubenswrapper[24928]: W1205 10:47:20.031907 24928 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:47:20.031931 master-0 kubenswrapper[24928]: W1205 10:47:20.031925 24928 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:47:20.031931 master-0 kubenswrapper[24928]: W1205 10:47:20.031930 24928 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031934 24928 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031938 24928 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031942 24928 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031946 24928 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031950 24928 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031954 24928 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031957 24928 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031961 24928 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031965 24928 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031968 24928 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031972 24928 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031981 24928 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031986 24928 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031990 24928 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031993 24928 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.031997 24928 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.032001 24928 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.032005 24928 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.032008 24928 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.032012 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:47:20.032028 master-0 kubenswrapper[24928]: W1205 10:47:20.032016 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032019 24928 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032023 24928 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032027 24928 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032030 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032034 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032037 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032041 24928 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032045 24928 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032048 24928 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032052 24928 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032056 24928 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032059 24928 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032063 24928 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032067 24928 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032070 24928 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032074 24928 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032078 24928 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032081 24928 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032085 24928 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:47:20.032670 master-0 kubenswrapper[24928]: W1205 10:47:20.032088 24928 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032092 24928 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032096 24928 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032099 24928 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032103 24928 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032106 24928 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032112 24928 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032117 24928 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032121 24928 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032125 24928 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032129 24928 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032133 24928 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032137 24928 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032142 24928 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032147 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032151 24928 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032156 24928 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032160 24928 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032164 24928 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:47:20.033156 master-0 kubenswrapper[24928]: W1205 10:47:20.032169 24928 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: W1205 10:47:20.032172 24928 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: W1205 10:47:20.032177 24928 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: W1205 10:47:20.032181 24928 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: W1205 10:47:20.032186 24928 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: W1205 10:47:20.032191 24928 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: W1205 10:47:20.032196 24928 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: W1205 10:47:20.032200 24928 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: W1205 10:47:20.032204 24928 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: W1205 10:47:20.032209 24928 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032280 24928 flags.go:64] FLAG: --address="0.0.0.0" Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032289 24928 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032296 24928 flags.go:64] FLAG: --anonymous-auth="true" Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032301 24928 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032306 24928 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032311 24928 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032316 24928 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032322 24928 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032326 24928 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032330 24928 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032335 24928 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 05 10:47:20.033692 master-0 kubenswrapper[24928]: I1205 10:47:20.032340 24928 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032344 24928 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032348 24928 flags.go:64] FLAG: --cgroup-root="" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032353 24928 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032358 24928 flags.go:64] FLAG: --client-ca-file="" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032362 24928 flags.go:64] FLAG: --cloud-config="" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032368 24928 flags.go:64] FLAG: --cloud-provider="" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032372 24928 flags.go:64] FLAG: --cluster-dns="[]" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032377 24928 flags.go:64] FLAG: --cluster-domain="" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032381 24928 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032386 24928 flags.go:64] FLAG: --config-dir="" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032390 24928 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032395 24928 flags.go:64] FLAG: --container-log-max-files="5" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032400 24928 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032404 24928 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032408 24928 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032412 24928 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032445 24928 flags.go:64] FLAG: --contention-profiling="false" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032450 24928 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032454 24928 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032459 24928 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032463 24928 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032469 24928 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032473 24928 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032477 24928 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 05 10:47:20.034250 master-0 kubenswrapper[24928]: I1205 10:47:20.032481 24928 flags.go:64] FLAG: --enable-load-reader="false" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032485 24928 flags.go:64] FLAG: --enable-server="true" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032490 24928 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032495 24928 flags.go:64] FLAG: --event-burst="100" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032500 24928 flags.go:64] FLAG: --event-qps="50" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032504 24928 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032509 24928 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032513 24928 flags.go:64] FLAG: --eviction-hard="" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032518 24928 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032522 24928 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032533 24928 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032538 24928 flags.go:64] FLAG: --eviction-soft="" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032542 24928 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032546 24928 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032551 24928 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032556 24928 flags.go:64] FLAG: --experimental-mounter-path="" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032560 24928 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032564 24928 flags.go:64] FLAG: --fail-swap-on="true" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032569 24928 flags.go:64] FLAG: --feature-gates="" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032574 24928 flags.go:64] FLAG: --file-check-frequency="20s" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032579 24928 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032583 24928 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032587 24928 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032592 24928 flags.go:64] FLAG: --healthz-port="10248" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032596 24928 flags.go:64] FLAG: --help="false" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032600 24928 flags.go:64] FLAG: --hostname-override="" Dec 05 10:47:20.034920 master-0 kubenswrapper[24928]: I1205 10:47:20.032604 24928 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032609 24928 flags.go:64] FLAG: --http-check-frequency="20s" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032613 24928 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032617 24928 flags.go:64] FLAG: --image-credential-provider-config="" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032621 24928 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032625 24928 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032629 24928 flags.go:64] FLAG: --image-service-endpoint="" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032634 24928 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032638 24928 flags.go:64] FLAG: --kube-api-burst="100" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032643 24928 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032647 24928 flags.go:64] FLAG: --kube-api-qps="50" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032652 24928 flags.go:64] FLAG: --kube-reserved="" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032656 24928 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032660 24928 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032664 24928 flags.go:64] FLAG: --kubelet-cgroups="" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032668 24928 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032675 24928 flags.go:64] FLAG: --lock-file="" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032679 24928 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032684 24928 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032688 24928 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032694 24928 flags.go:64] FLAG: --log-json-split-stream="false" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032699 24928 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032703 24928 flags.go:64] FLAG: --log-text-split-stream="false" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032708 24928 flags.go:64] FLAG: --logging-format="text" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032713 24928 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 05 10:47:20.035594 master-0 kubenswrapper[24928]: I1205 10:47:20.032717 24928 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032722 24928 flags.go:64] FLAG: --manifest-url="" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032726 24928 flags.go:64] FLAG: --manifest-url-header="" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032731 24928 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032736 24928 flags.go:64] FLAG: --max-open-files="1000000" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032741 24928 flags.go:64] FLAG: --max-pods="110" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032745 24928 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032749 24928 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032754 24928 flags.go:64] FLAG: --memory-manager-policy="None" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032758 24928 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032763 24928 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032768 24928 flags.go:64] FLAG: --node-ip="192.168.32.10" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032772 24928 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032781 24928 flags.go:64] FLAG: --node-status-max-images="50" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032785 24928 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032789 24928 flags.go:64] FLAG: --oom-score-adj="-999" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032794 24928 flags.go:64] FLAG: --pod-cidr="" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032798 24928 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a70b2a95140d1e90978f36cc9889013ae34bd232662c5424002274385669ed9" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032804 24928 flags.go:64] FLAG: --pod-manifest-path="" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032809 24928 flags.go:64] FLAG: --pod-max-pids="-1" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032813 24928 flags.go:64] FLAG: --pods-per-core="0" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032817 24928 flags.go:64] FLAG: --port="10250" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032822 24928 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032826 24928 flags.go:64] FLAG: --provider-id="" Dec 05 10:47:20.036194 master-0 kubenswrapper[24928]: I1205 10:47:20.032830 24928 flags.go:64] FLAG: --qos-reserved="" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032835 24928 flags.go:64] FLAG: --read-only-port="10255" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032839 24928 flags.go:64] FLAG: --register-node="true" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032844 24928 flags.go:64] FLAG: --register-schedulable="true" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032848 24928 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032855 24928 flags.go:64] FLAG: --registry-burst="10" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032859 24928 flags.go:64] FLAG: --registry-qps="5" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032863 24928 flags.go:64] FLAG: --reserved-cpus="" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032867 24928 flags.go:64] FLAG: --reserved-memory="" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032872 24928 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032877 24928 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032881 24928 flags.go:64] FLAG: --rotate-certificates="false" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032886 24928 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032891 24928 flags.go:64] FLAG: --runonce="false" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032895 24928 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032899 24928 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032903 24928 flags.go:64] FLAG: --seccomp-default="false" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032908 24928 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032912 24928 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032916 24928 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032921 24928 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032925 24928 flags.go:64] FLAG: --storage-driver-password="root" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032929 24928 flags.go:64] FLAG: --storage-driver-secure="false" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032933 24928 flags.go:64] FLAG: --storage-driver-table="stats" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032937 24928 flags.go:64] FLAG: --storage-driver-user="root" Dec 05 10:47:20.037409 master-0 kubenswrapper[24928]: I1205 10:47:20.032942 24928 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032946 24928 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032951 24928 flags.go:64] FLAG: --system-cgroups="" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032955 24928 flags.go:64] FLAG: --system-reserved="cpu=500m,ephemeral-storage=1Gi,memory=1Gi" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032961 24928 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032966 24928 flags.go:64] FLAG: --tls-cert-file="" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032970 24928 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032975 24928 flags.go:64] FLAG: --tls-min-version="" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032979 24928 flags.go:64] FLAG: --tls-private-key-file="" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032984 24928 flags.go:64] FLAG: --topology-manager-policy="none" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032988 24928 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032992 24928 flags.go:64] FLAG: --topology-manager-scope="container" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.032996 24928 flags.go:64] FLAG: --v="2" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.033001 24928 flags.go:64] FLAG: --version="false" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.033008 24928 flags.go:64] FLAG: --vmodule="" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.033013 24928 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: I1205 10:47:20.033017 24928 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: W1205 10:47:20.033110 24928 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: W1205 10:47:20.033115 24928 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: W1205 10:47:20.033120 24928 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: W1205 10:47:20.033124 24928 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: W1205 10:47:20.033128 24928 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: W1205 10:47:20.033132 24928 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:47:20.038774 master-0 kubenswrapper[24928]: W1205 10:47:20.033136 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033140 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033144 24928 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033148 24928 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033153 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033157 24928 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033160 24928 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033165 24928 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033168 24928 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033172 24928 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033176 24928 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033179 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033183 24928 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033188 24928 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033193 24928 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033197 24928 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033201 24928 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033204 24928 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033208 24928 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:47:20.039727 master-0 kubenswrapper[24928]: W1205 10:47:20.033212 24928 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033216 24928 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033221 24928 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033226 24928 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033230 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033235 24928 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033239 24928 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033243 24928 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033247 24928 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033251 24928 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033254 24928 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033258 24928 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033262 24928 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033265 24928 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033269 24928 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033273 24928 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033278 24928 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033281 24928 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033285 24928 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:47:20.040833 master-0 kubenswrapper[24928]: W1205 10:47:20.033289 24928 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033293 24928 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033297 24928 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033301 24928 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033304 24928 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033308 24928 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033312 24928 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033318 24928 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033322 24928 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033325 24928 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033329 24928 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033333 24928 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033337 24928 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033341 24928 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033345 24928 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033349 24928 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033352 24928 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033356 24928 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033360 24928 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033364 24928 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033367 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:47:20.041949 master-0 kubenswrapper[24928]: W1205 10:47:20.033371 24928 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.033375 24928 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.033378 24928 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.033382 24928 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.033386 24928 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.033390 24928 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.033393 24928 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: I1205 10:47:20.033400 24928 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false StreamingCollectionEncodingToJSON:false StreamingCollectionEncodingToProtobuf:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: I1205 10:47:20.040732 24928 server.go:491] "Kubelet version" kubeletVersion="v1.31.13" Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: I1205 10:47:20.040783 24928 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.040969 24928 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.040983 24928 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.040990 24928 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.041013 24928 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.041019 24928 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:47:20.043599 master-0 kubenswrapper[24928]: W1205 10:47:20.041024 24928 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041033 24928 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041038 24928 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041043 24928 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041067 24928 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041074 24928 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041079 24928 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041083 24928 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041090 24928 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041099 24928 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041105 24928 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041110 24928 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041115 24928 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041120 24928 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041128 24928 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041132 24928 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041137 24928 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041142 24928 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041147 24928 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041152 24928 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:47:20.044280 master-0 kubenswrapper[24928]: W1205 10:47:20.041157 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041163 24928 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041179 24928 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041185 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041189 24928 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041193 24928 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041197 24928 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041203 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041207 24928 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041211 24928 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041216 24928 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041221 24928 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041226 24928 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041231 24928 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041235 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041238 24928 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041242 24928 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041246 24928 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041250 24928 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041257 24928 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:47:20.045457 master-0 kubenswrapper[24928]: W1205 10:47:20.041271 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041276 24928 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041281 24928 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041286 24928 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041290 24928 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041295 24928 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041300 24928 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041304 24928 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041309 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041314 24928 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041317 24928 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041324 24928 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041329 24928 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041332 24928 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041350 24928 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041355 24928 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041359 24928 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041364 24928 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041371 24928 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:47:20.046238 master-0 kubenswrapper[24928]: W1205 10:47:20.041376 24928 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.041382 24928 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.041387 24928 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.041391 24928 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.041401 24928 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.041406 24928 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.041410 24928 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.041414 24928 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: I1205 10:47:20.041438 24928 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false StreamingCollectionEncodingToJSON:false StreamingCollectionEncodingToProtobuf:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.042055 24928 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.042073 24928 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.042080 24928 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.042086 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.042091 24928 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 10:47:20.047301 master-0 kubenswrapper[24928]: W1205 10:47:20.042096 24928 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042101 24928 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042105 24928 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042109 24928 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042113 24928 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042117 24928 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042121 24928 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042125 24928 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042129 24928 feature_gate.go:330] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042133 24928 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042136 24928 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042140 24928 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042144 24928 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042148 24928 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042152 24928 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042157 24928 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042161 24928 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042165 24928 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042168 24928 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042172 24928 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 10:47:20.048233 master-0 kubenswrapper[24928]: W1205 10:47:20.042176 24928 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042180 24928 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042183 24928 feature_gate.go:330] unrecognized feature gate: Example Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042187 24928 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042191 24928 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042194 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042198 24928 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042203 24928 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042207 24928 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042211 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042215 24928 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042220 24928 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042223 24928 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042227 24928 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042231 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042235 24928 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042239 24928 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042243 24928 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042247 24928 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042251 24928 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 10:47:20.049009 master-0 kubenswrapper[24928]: W1205 10:47:20.042255 24928 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042259 24928 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042262 24928 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042267 24928 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042272 24928 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042276 24928 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042281 24928 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042286 24928 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042292 24928 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042296 24928 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042300 24928 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042304 24928 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042309 24928 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042313 24928 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042318 24928 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042322 24928 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042325 24928 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042330 24928 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042334 24928 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 10:47:20.049802 master-0 kubenswrapper[24928]: W1205 10:47:20.042338 24928 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: W1205 10:47:20.042342 24928 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: W1205 10:47:20.042346 24928 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: W1205 10:47:20.042351 24928 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: W1205 10:47:20.042356 24928 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: W1205 10:47:20.042360 24928 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: W1205 10:47:20.042363 24928 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: W1205 10:47:20.042368 24928 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: I1205 10:47:20.042376 24928 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false StreamingCollectionEncodingToJSON:false StreamingCollectionEncodingToProtobuf:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: I1205 10:47:20.042602 24928 server.go:940] "Client rotation is on, will bootstrap in background" Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: I1205 10:47:20.044223 24928 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: I1205 10:47:20.044300 24928 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: I1205 10:47:20.044526 24928 server.go:997] "Starting client certificate rotation" Dec 05 10:47:20.050863 master-0 kubenswrapper[24928]: I1205 10:47:20.044537 24928 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 05 10:47:20.051613 master-0 kubenswrapper[24928]: I1205 10:47:20.044737 24928 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2025-12-06 10:27:25 +0000 UTC, rotation deadline is 2025-12-06 03:26:16.911930438 +0000 UTC Dec 05 10:47:20.051613 master-0 kubenswrapper[24928]: I1205 10:47:20.044834 24928 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 16h38m56.867102229s for next certificate rotation Dec 05 10:47:20.051613 master-0 kubenswrapper[24928]: I1205 10:47:20.045247 24928 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 10:47:20.051613 master-0 kubenswrapper[24928]: I1205 10:47:20.046480 24928 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 10:47:20.051613 master-0 kubenswrapper[24928]: I1205 10:47:20.050709 24928 log.go:25] "Validated CRI v1 runtime API" Dec 05 10:47:20.056492 master-0 kubenswrapper[24928]: I1205 10:47:20.056366 24928 log.go:25] "Validated CRI v1 image API" Dec 05 10:47:20.058202 master-0 kubenswrapper[24928]: I1205 10:47:20.057650 24928 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 05 10:47:20.071261 master-0 kubenswrapper[24928]: I1205 10:47:20.071167 24928 fs.go:135] Filesystem UUIDs: map[354073b4-663b-4c82-b811-53676c745317:/dev/vda3 7B77-95E7:/dev/vda2 910678ff-f77e-4a7d-8d53-86f2ac47a823:/dev/vda4] Dec 05 10:47:20.072277 master-0 kubenswrapper[24928]: I1205 10:47:20.071247 24928 fs.go:136] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/07cb2a5b59f4a1cf9ae60c0a62bf4e44ad34c98cbd92efe399d17d750ec292c6/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/07cb2a5b59f4a1cf9ae60c0a62bf4e44ad34c98cbd92efe399d17d750ec292c6/userdata/shm major:0 minor:1176 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/088f93fb71215e03078ad676d01b6c7b68c855017eff77ac0edbcdeaa645fd6c/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/088f93fb71215e03078ad676d01b6c7b68c855017eff77ac0edbcdeaa645fd6c/userdata/shm major:0 minor:503 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/0d2211dcab91af9edc388c398c5fc7a90c2e8c65270b4ebc9dead574b96eaefe/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/0d2211dcab91af9edc388c398c5fc7a90c2e8c65270b4ebc9dead574b96eaefe/userdata/shm major:0 minor:332 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/0d874913b26ce96decb20e6d175c667c355d6a65334e625b074f3811b11e3027/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/0d874913b26ce96decb20e6d175c667c355d6a65334e625b074f3811b11e3027/userdata/shm major:0 minor:541 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/0e02576ddbd856f2a1c30d4bbdcb0b27384c73dfe012be69bbb5c2168f8df7fd/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/0e02576ddbd856f2a1c30d4bbdcb0b27384c73dfe012be69bbb5c2168f8df7fd/userdata/shm major:0 minor:621 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/0e8ac47ff99876f6c126e479233a75b5de6a9aabecbf3b414116156cf0e23f86/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/0e8ac47ff99876f6c126e479233a75b5de6a9aabecbf3b414116156cf0e23f86/userdata/shm major:0 minor:894 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/0f19c2361e4856d55a1f7cfcf2bbf2e6aae6be0b4b861e0e68ba7772ccbc2487/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/0f19c2361e4856d55a1f7cfcf2bbf2e6aae6be0b4b861e0e68ba7772ccbc2487/userdata/shm major:0 minor:875 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/12d79e8d025da0e5c89728d8b35230c8c9764259d39417d2ff38f00dc9c263f6/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/12d79e8d025da0e5c89728d8b35230c8c9764259d39417d2ff38f00dc9c263f6/userdata/shm major:0 minor:349 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/1bde7edcbb5c913afcdfed1e9b165ffe883b909863e89e771b54560c18c0ac2a/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/1bde7edcbb5c913afcdfed1e9b165ffe883b909863e89e771b54560c18c0ac2a/userdata/shm major:0 minor:954 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/2421d54e72b15809966ab160149e8662f98357fd3da518130cc175e2f9f3791a/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/2421d54e72b15809966ab160149e8662f98357fd3da518130cc175e2f9f3791a/userdata/shm major:0 minor:1446 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/2543decbaf3eadc4ac21f8d8b5c5c326c025c08f080ed2e80c25c76a6e2b34dd/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/2543decbaf3eadc4ac21f8d8b5c5c326c025c08f080ed2e80c25c76a6e2b34dd/userdata/shm major:0 minor:1638 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/2c8d17056f01fe2b63601e7d3f2b689f6a1b5945d124ed1dd3ebcdd26090636d/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/2c8d17056f01fe2b63601e7d3f2b689f6a1b5945d124ed1dd3ebcdd26090636d/userdata/shm major:0 minor:534 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/2cf0d485e10f74fdd9bf0282b8f6fbc341fecfb79764618939e754565cc0b6ff/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/2cf0d485e10f74fdd9bf0282b8f6fbc341fecfb79764618939e754565cc0b6ff/userdata/shm major:0 minor:1076 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/35b4f066f81dc683769e8a58a956f46b65324a32b3270c15056ef67e541a8fdd/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/35b4f066f81dc683769e8a58a956f46b65324a32b3270c15056ef67e541a8fdd/userdata/shm major:0 minor:1175 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/4446e01522dc26b79f498f8e0dc137457645ab081cac6529b44e9b557b9d72f8/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/4446e01522dc26b79f498f8e0dc137457645ab081cac6529b44e9b557b9d72f8/userdata/shm major:0 minor:136 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/45e53c8beb8eb2f205541b531720f581844942fc2ea4f6200401dc7c6d748d82/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/45e53c8beb8eb2f205541b531720f581844942fc2ea4f6200401dc7c6d748d82/userdata/shm major:0 minor:1312 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/4833f667ac4451a150ed477789be72c08b65162d8c7bbf29e44cd656299dbd63/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/4833f667ac4451a150ed477789be72c08b65162d8c7bbf29e44cd656299dbd63/userdata/shm major:0 minor:375 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/487c60a3656ecce96053a1bf83cb0f4b361d27909df0d101db089d62b7524b7b/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/487c60a3656ecce96053a1bf83cb0f4b361d27909df0d101db089d62b7524b7b/userdata/shm major:0 minor:619 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/4c76b1c660e01ccc1ed121db81862ca5648ff5a628cbb72bf8b46891908897a5/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/4c76b1c660e01ccc1ed121db81862ca5648ff5a628cbb72bf8b46891908897a5/userdata/shm major:0 minor:1172 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/4cedf7a9bbcb2d72fff4df0c2eb3dda7d05531d0220b30c3f687e88816b7ce46/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/4cedf7a9bbcb2d72fff4df0c2eb3dda7d05531d0220b30c3f687e88816b7ce46/userdata/shm major:0 minor:1478 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/4f32a485a101f0d25f8273536aa5f41f281533450b1891e0d6dbaa0dcf7b996e/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/4f32a485a101f0d25f8273536aa5f41f281533450b1891e0d6dbaa0dcf7b996e/userdata/shm major:0 minor:1474 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/5046caa4441a1760a91554be78c89b5b216c5aae3f10502f03a571a1cc07afd4/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/5046caa4441a1760a91554be78c89b5b216c5aae3f10502f03a571a1cc07afd4/userdata/shm major:0 minor:1069 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/52b24a28414c04f2574070f038561f387ec44118f13f6cac5eab1761c0750f3a/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/52b24a28414c04f2574070f038561f387ec44118f13f6cac5eab1761c0750f3a/userdata/shm major:0 minor:1173 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/59a797eb008b1a85db4180d24d857e8f08956acdd19cbd4d960f8adc3117a02b/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/59a797eb008b1a85db4180d24d857e8f08956acdd19cbd4d960f8adc3117a02b/userdata/shm major:0 minor:555 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/5ab705ed0c0dfea5ee8845410f03f5084071123b3566b80a7ef7e6fcce0fa974/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/5ab705ed0c0dfea5ee8845410f03f5084071123b3566b80a7ef7e6fcce0fa974/userdata/shm major:0 minor:383 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/5b905a3a36c591b5f3704d14e85aafe47726c309b50985c9308b5ed0d4a90267/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/5b905a3a36c591b5f3704d14e85aafe47726c309b50985c9308b5ed0d4a90267/userdata/shm major:0 minor:346 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/5be78d3569389b67ef6d6c6575e7c97a75617917f6e7f31b9b05a692b9e12292/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/5be78d3569389b67ef6d6c6575e7c97a75617917f6e7f31b9b05a692b9e12292/userdata/shm major:0 minor:55 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/5cd120c4a52499689b34a1d9baed806acbb93f12cd8a2bcbf5cdb4f6c23214df/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/5cd120c4a52499689b34a1d9baed806acbb93f12cd8a2bcbf5cdb4f6c23214df/userdata/shm major:0 minor:1344 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/6010e9394b879952b424eb35b6227a92d6319988a7bfc5f908a8adc470e9463b/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/6010e9394b879952b424eb35b6227a92d6319988a7bfc5f908a8adc470e9463b/userdata/shm major:0 minor:1077 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/61b925377e25c3f9cf64e2a949ea1fe86f2ea5c4e951a4e7adc2aa3b0cb3336d/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/61b925377e25c3f9cf64e2a949ea1fe86f2ea5c4e951a4e7adc2aa3b0cb3336d/userdata/shm major:0 minor:1343 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/62e34bf231128449a630a3dad98495dd55715547e65586a815afe200566d05a0/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/62e34bf231128449a630a3dad98495dd55715547e65586a815afe200566d05a0/userdata/shm major:0 minor:622 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/6607da6ec4214a3fe2d315e70d9756b15e987231bb8310517d849da88c8e7b15/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/6607da6ec4214a3fe2d315e70d9756b15e987231bb8310517d849da88c8e7b15/userdata/shm major:0 minor:165 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/6ead10efe7a643d9b0dd883cc19f8ef852a5658bc79c235b7c7c6b5de2e97811/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/6ead10efe7a643d9b0dd883cc19f8ef852a5658bc79c235b7c7c6b5de2e97811/userdata/shm major:0 minor:189 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/70e7f53d39e81bfad8819ec3c2244577ab77bdd0b9a5512ba3a7cb61ae8bafb4/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/70e7f53d39e81bfad8819ec3c2244577ab77bdd0b9a5512ba3a7cb61ae8bafb4/userdata/shm major:0 minor:473 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/77b83e99ece1e07058ac092ebb18d76d85bb5a90292440cd154e967f63253776/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/77b83e99ece1e07058ac092ebb18d76d85bb5a90292440cd154e967f63253776/userdata/shm major:0 minor:626 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/7c5339e633ef9bd66b822a691a1a94b12e50803e5297f66f5662bfb59a1a2a6c/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/7c5339e633ef9bd66b822a691a1a94b12e50803e5297f66f5662bfb59a1a2a6c/userdata/shm major:0 minor:1072 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/7e22b55ec9c49898cca9f69ea15552e8a230cbb6781fbf80f0e4919b22b7b7d8/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/7e22b55ec9c49898cca9f69ea15552e8a230cbb6781fbf80f0e4919b22b7b7d8/userdata/shm major:0 minor:338 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/7f2b0f7328f1fdf55c43aa55c01436ada2488903a3cf2d212fd3c7469222fb7b/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/7f2b0f7328f1fdf55c43aa55c01436ada2488903a3cf2d212fd3c7469222fb7b/userdata/shm major:0 minor:1024 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/8066239534fa93e4ea7b428954f42f9cdca327545a8ae0d3c647f8ee2fd24e03/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/8066239534fa93e4ea7b428954f42f9cdca327545a8ae0d3c647f8ee2fd24e03/userdata/shm major:0 minor:531 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/80b1488f509c166b5b116f44dccb9102304f263ee75bb4a43991cef27c102c84/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/80b1488f509c166b5b116f44dccb9102304f263ee75bb4a43991cef27c102c84/userdata/shm major:0 minor:532 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/8114256db0301425275a52e5c6f4316fb7a7d5eed5146413b166f9ab3918f143/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/8114256db0301425275a52e5c6f4316fb7a7d5eed5146413b166f9ab3918f143/userdata/shm major:0 minor:341 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/812553092d13019021bace1ffb05fbafd76c83d356691f1b938f38559234f787/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/812553092d13019021bace1ffb05fbafd76c83d356691f1b938f38559234f787/userdata/shm major:0 minor:1340 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/8904f4db7f08be6823cd5712612b6b29bb4d4506ff20a7f693a2b99016d20918/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/8904f4db7f08be6823cd5712612b6b29bb4d4506ff20a7f693a2b99016d20918/userdata/shm major:0 minor:1177 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/897152fdb47d25f0a5016944193be33416b9e61d9cbcab904f0e8c6b63de2a36/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/897152fdb47d25f0a5016944193be33416b9e61d9cbcab904f0e8c6b63de2a36/userdata/shm major:0 minor:1062 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/89c8e517113e00942d58576e87acb9fd6a3e8f726e30d612913a3d827c50d10f/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/89c8e517113e00942d58576e87acb9fd6a3e8f726e30d612913a3d827c50d10f/userdata/shm major:0 minor:485 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/8d62f303e35a44bd684cd9b7f51835e4e80959496c433fcc66b39cdd84c84cc8/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/8d62f303e35a44bd684cd9b7f51835e4e80959496c433fcc66b39cdd84c84cc8/userdata/shm major:0 minor:149 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/97c0ac3e4d477fff0dd9953d178193fc87456dca36efdf5e88346f7b00aeff5b/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/97c0ac3e4d477fff0dd9953d178193fc87456dca36efdf5e88346f7b00aeff5b/userdata/shm major:0 minor:525 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/97d00d6aa4d6661e2be349a4e20636f067fc69912c7110a63d99e820d5970ef4/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/97d00d6aa4d6661e2be349a4e20636f067fc69912c7110a63d99e820d5970ef4/userdata/shm major:0 minor:718 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/9dab351b1866858c81dfc4a1e5ed04a7dcaa8ee50017ec3f0b131de9c613ce86/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/9dab351b1866858c81dfc4a1e5ed04a7dcaa8ee50017ec3f0b131de9c613ce86/userdata/shm major:0 minor:1560 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/9e8deb0a79a04a088af836af85742626a8778be7f61d22448e3a0b846ff8b05a/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/9e8deb0a79a04a088af836af85742626a8778be7f61d22448e3a0b846ff8b05a/userdata/shm major:0 minor:1004 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/a669846f3b9be1321aadc153fc634864e5c4e6b7c212feee982c065424fad087/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/a669846f3b9be1321aadc153fc634864e5c4e6b7c212feee982c065424fad087/userdata/shm major:0 minor:533 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/aa4c28f3537eda89f3c42c0956c1461303fba100316bf76337e7c6ab1fb2e2e4/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/aa4c28f3537eda89f3c42c0956c1461303fba100316bf76337e7c6ab1fb2e2e4/userdata/shm major:0 minor:639 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/aa871dfa794fea5d166618a43475fe935ef7d4d3799afbc3d92debffe4b76081/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/aa871dfa794fea5d166618a43475fe935ef7d4d3799afbc3d92debffe4b76081/userdata/shm major:0 minor:623 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/ad7983dc1d046a27e0480393c3ce354f96d0968b079cf24fa1ebf196a469d66d/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/ad7983dc1d046a27e0480393c3ce354f96d0968b079cf24fa1ebf196a469d66d/userdata/shm major:0 minor:127 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/b0fb01570845c0cbcb28b2232b0ade7593b0c0de030669cae5f3ef53544cda81/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/b0fb01570845c0cbcb28b2232b0ade7593b0c0de030669cae5f3ef53544cda81/userdata/shm major:0 minor:1056 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/b44ac27d1b7f531e11c3f4d352d89c4a7679a795ec64417e43c87c5d1e725e90/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/b44ac27d1b7f531e11c3f4d352d89c4a7679a795ec64417e43c87c5d1e725e90/userdata/shm major:0 minor:314 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/b73ea430e90c45d058f4dc00766a5e44931feca970feb8976e3fca4580bc112f/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/b73ea430e90c45d058f4dc00766a5e44931feca970feb8976e3fca4580bc112f/userdata/shm major:0 minor:95 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/b7452830de99c0487a8d247b23da92391c3a1113eb379e2b40ac1673b8cb1fe1/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/b7452830de99c0487a8d247b23da92391c3a1113eb379e2b40ac1673b8cb1fe1/userdata/shm major:0 minor:336 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/b94fba3ebe8eacc577514add51b89ec37946322bc96bbe0ee6d3e8d5b7830ae3/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/b94fba3ebe8eacc577514add51b89ec37946322bc96bbe0ee6d3e8d5b7830ae3/userdata/shm major:0 minor:1369 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/ba6a43bb61d242c8e032f7cab8a9cf0e806cdf03c0a9ce44ec6c444a2216f39a/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/ba6a43bb61d242c8e032f7cab8a9cf0e806cdf03c0a9ce44ec6c444a2216f39a/userdata/shm major:0 minor:624 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/bd5095616b6a8494a79971c731d5ef49c3bae5cbd4111190f1e5dfde0e3d9554/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/bd5095616b6a8494a79971c731d5ef49c3bae5cbd4111190f1e5dfde0e3d9554/userdata/shm major:0 minor:812 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/bdfbd9daf37e036f9b2d20314905a9aa48f68e21954c4e133ee980fe10b0d8d2/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/bdfbd9daf37e036f9b2d20314905a9aa48f68e21954c4e133ee980fe10b0d8d2/userdata/shm major:0 minor:343 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/c22e02aa77c0fa78cb950ff2c4f24a8244e5ac70b15a797f17b5f262893f0220/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/c22e02aa77c0fa78cb950ff2c4f24a8244e5ac70b15a797f17b5f262893f0220/userdata/shm major:0 minor:1068 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/c93da22a302a383f4fdc18e25cf772f0d9d551ca131191ed837accfc6f610c6f/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/c93da22a302a383f4fdc18e25cf772f0d9d551ca131191ed837accfc6f610c6f/userdata/shm major:0 minor:1064 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/ccbec542fddb9bd0831349589c4ac65650d79b1cf75a0af75cf8ca2d1a0d89f6/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/ccbec542fddb9bd0831349589c4ac65650d79b1cf75a0af75cf8ca2d1a0d89f6/userdata/shm major:0 minor:337 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7/userdata/shm major:0 minor:41 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/d22a14317ab6ff7dd171c6dc6d6a0e336ee5087a9085305232d6accf32502f02/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/d22a14317ab6ff7dd171c6dc6d6a0e336ee5087a9085305232d6accf32502f02/userdata/shm major:0 minor:54 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/d4750111f1d0e54879f82e850c1814dfc68424e9b8c03f07177ca16103b8a79d/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/d4750111f1d0e54879f82e850c1814dfc68424e9b8c03f07177ca16103b8a79d/userdata/shm major:0 minor:1066 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/d996cd9dee5c72648a9d435634d8c36eb7feb019522350be8d7d8f4a1fa79c46/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/d996cd9dee5c72648a9d435634d8c36eb7feb019522350be8d7d8f4a1fa79c46/userdata/shm major:0 minor:1482 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/e2506cc6607d77e49135fcb375a37bafe2d511f6ed75718677873e40b092ab87/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/e2506cc6607d77e49135fcb375a37bafe2d511f6ed75718677873e40b092ab87/userdata/shm major:0 minor:333 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/e88afc2116e69a8053c275a20f5cc4b6af9c5241405fc2bf27810f224b5549fb/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/e88afc2116e69a8053c275a20f5cc4b6af9c5241405fc2bf27810f224b5549fb/userdata/shm major:0 minor:1146 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/eecf6d3a967ce077414ad07bec5a21f5acf63791b8f14de4d8d34c6cfe91444a/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/eecf6d3a967ce077414ad07bec5a21f5acf63791b8f14de4d8d34c6cfe91444a/userdata/shm major:0 minor:468 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/efc2234f689663f4f73596bfe5f7a66235fb095a2337fa0c8bc412e1e08433fb/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/efc2234f689663f4f73596bfe5f7a66235fb095a2337fa0c8bc412e1e08433fb/userdata/shm major:0 minor:139 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/f12647ed460372c13fb857a5856ac53a5f123d196cc70548c9f47e2b35c490ce/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/f12647ed460372c13fb857a5856ac53a5f123d196cc70548c9f47e2b35c490ce/userdata/shm major:0 minor:1060 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/f77a8c44878083fb1f5b837847555bd296fce89d49bff4c5ba5fc4a9953922c2/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/f77a8c44878083fb1f5b837847555bd296fce89d49bff4c5ba5fc4a9953922c2/userdata/shm major:0 minor:1073 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/fc75409854b1370368a39cd91bcd5c4ac2aa319c2a8b9aaad1e7abab3fd56a49/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/fc75409854b1370368a39cd91bcd5c4ac2aa319c2a8b9aaad1e7abab3fd56a49/userdata/shm major:0 minor:179 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/fec7b5fa2a30a341bff8e373ec4fbcd19900103be9e9af536060bcabaccdcce1/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/fec7b5fa2a30a341bff8e373ec4fbcd19900103be9e9af536060bcabaccdcce1/userdata/shm major:0 minor:727 fsType:tmpfs blockSize:0} /run/containers/storage/overlay-containers/fedfe3c8bf160c89d4d691575ab31b377473c05a2dd753786d2ef1a80ea5d562/userdata/shm:{mountpoint:/run/containers/storage/overlay-containers/fedfe3c8bf160c89d4d691575ab31b377473c05a2dd753786d2ef1a80ea5d562/userdata/shm major:0 minor:347 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/01488ac4-313d-48d5-9e86-7b2011b9e91e/volumes/kubernetes.io~projected/kube-api-access-2xjx8:{mountpoint:/var/lib/kubelet/pods/01488ac4-313d-48d5-9e86-7b2011b9e91e/volumes/kubernetes.io~projected/kube-api-access-2xjx8 major:0 minor:1048 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/01488ac4-313d-48d5-9e86-7b2011b9e91e/volumes/kubernetes.io~secret/proxy-tls:{mountpoint:/var/lib/kubelet/pods/01488ac4-313d-48d5-9e86-7b2011b9e91e/volumes/kubernetes.io~secret/proxy-tls major:0 minor:1044 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/0a4db531-8af7-4085-a6b2-0de51b527ce6/volumes/kubernetes.io~projected/kube-api-access-g4jgc:{mountpoint:/var/lib/kubelet/pods/0a4db531-8af7-4085-a6b2-0de51b527ce6/volumes/kubernetes.io~projected/kube-api-access-g4jgc major:0 minor:1171 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/0aa6e138-4b1f-4047-8255-a2b14d044588/volumes/kubernetes.io~projected/kube-api-access-wnzgs:{mountpoint:/var/lib/kubelet/pods/0aa6e138-4b1f-4047-8255-a2b14d044588/volumes/kubernetes.io~projected/kube-api-access-wnzgs major:0 minor:1168 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/0babac16-164a-405a-a86c-30524118f046/volumes/kubernetes.io~secret/tls-certificates:{mountpoint:/var/lib/kubelet/pods/0babac16-164a-405a-a86c-30524118f046/volumes/kubernetes.io~secret/tls-certificates major:0 minor:1337 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/0de26264-46c2-4d29-97da-25a1681d6a8e/volumes/kubernetes.io~projected/kube-api-access-cdc7g:{mountpoint:/var/lib/kubelet/pods/0de26264-46c2-4d29-97da-25a1681d6a8e/volumes/kubernetes.io~projected/kube-api-access-cdc7g major:0 minor:1050 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/0de26264-46c2-4d29-97da-25a1681d6a8e/volumes/kubernetes.io~secret/samples-operator-tls:{mountpoint:/var/lib/kubelet/pods/0de26264-46c2-4d29-97da-25a1681d6a8e/volumes/kubernetes.io~secret/samples-operator-tls major:0 minor:1037 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~projected/kube-api-access:{mountpoint:/var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~projected/kube-api-access major:0 minor:281 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~secret/serving-cert major:0 minor:256 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~projected/kube-api-access-cwqkb:{mountpoint:/var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~projected/kube-api-access-cwqkb major:0 minor:122 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~secret/metrics-tls:{mountpoint:/var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~secret/metrics-tls major:0 minor:43 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~projected/kube-api-access-7wfsv:{mountpoint:/var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~projected/kube-api-access-7wfsv major:0 minor:282 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~secret/serving-cert major:0 minor:257 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1f132702-304b-46ff-b428-6df1deeffec3/volumes/kubernetes.io~projected/kube-api-access-vkzkh:{mountpoint:/var/lib/kubelet/pods/1f132702-304b-46ff-b428-6df1deeffec3/volumes/kubernetes.io~projected/kube-api-access-vkzkh major:0 minor:1049 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1f132702-304b-46ff-b428-6df1deeffec3/volumes/kubernetes.io~secret/cert:{mountpoint:/var/lib/kubelet/pods/1f132702-304b-46ff-b428-6df1deeffec3/volumes/kubernetes.io~secret/cert major:0 minor:1043 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1fd0a349-0f66-4c85-95e4-8bbc96648c05/volumes/kubernetes.io~projected/kube-api-access:{mountpoint:/var/lib/kubelet/pods/1fd0a349-0f66-4c85-95e4-8bbc96648c05/volumes/kubernetes.io~projected/kube-api-access major:0 minor:884 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/1fd0a349-0f66-4c85-95e4-8bbc96648c05/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/1fd0a349-0f66-4c85-95e4-8bbc96648c05/volumes/kubernetes.io~secret/serving-cert major:0 minor:883 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/21a9bd5a-3ade-48ef-8004-e0492daa85ba/volumes/kubernetes.io~projected/kube-api-access-r8bpw:{mountpoint:/var/lib/kubelet/pods/21a9bd5a-3ade-48ef-8004-e0492daa85ba/volumes/kubernetes.io~projected/kube-api-access-r8bpw major:0 minor:1476 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/21a9bd5a-3ade-48ef-8004-e0492daa85ba/volumes/kubernetes.io~secret/node-exporter-kube-rbac-proxy-config:{mountpoint:/var/lib/kubelet/pods/21a9bd5a-3ade-48ef-8004-e0492daa85ba/volumes/kubernetes.io~secret/node-exporter-kube-rbac-proxy-config major:0 minor:1471 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/21a9bd5a-3ade-48ef-8004-e0492daa85ba/volumes/kubernetes.io~secret/node-exporter-tls:{mountpoint:/var/lib/kubelet/pods/21a9bd5a-3ade-48ef-8004-e0492daa85ba/volumes/kubernetes.io~secret/node-exporter-tls major:0 minor:1470 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/bound-sa-token:{mountpoint:/var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/bound-sa-token major:0 minor:308 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/kube-api-access-vkqz7:{mountpoint:/var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/kube-api-access-vkqz7 major:0 minor:287 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~secret/metrics-tls:{mountpoint:/var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~secret/metrics-tls major:0 minor:556 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/24444474-aa9a-4a0f-8b4d-90f0009e0dc7/volumes/kubernetes.io~projected/kube-api-access-lrwt6:{mountpoint:/var/lib/kubelet/pods/24444474-aa9a-4a0f-8b4d-90f0009e0dc7/volumes/kubernetes.io~projected/kube-api-access-lrwt6 major:0 minor:690 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/24444474-aa9a-4a0f-8b4d-90f0009e0dc7/volumes/kubernetes.io~secret/control-plane-machine-set-operator-tls:{mountpoint:/var/lib/kubelet/pods/24444474-aa9a-4a0f-8b4d-90f0009e0dc7/volumes/kubernetes.io~secret/control-plane-machine-set-operator-tls major:0 minor:689 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/283122ba-be1c-4516-bd0f-df41c13c098b/volumes/kubernetes.io~projected/kube-api-access-vhjhk:{mountpoint:/var/lib/kubelet/pods/283122ba-be1c-4516-bd0f-df41c13c098b/volumes/kubernetes.io~projected/kube-api-access-vhjhk major:0 minor:275 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/2c3fab4d-05b8-42ec-8c91-91ff64562649/volumes/kubernetes.io~projected/kube-api-access-l8wnp:{mountpoint:/var/lib/kubelet/pods/2c3fab4d-05b8-42ec-8c91-91ff64562649/volumes/kubernetes.io~projected/kube-api-access-l8wnp major:0 minor:1051 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/2c3fab4d-05b8-42ec-8c91-91ff64562649/volumes/kubernetes.io~secret/apiservice-cert:{mountpoint:/var/lib/kubelet/pods/2c3fab4d-05b8-42ec-8c91-91ff64562649/volumes/kubernetes.io~secret/apiservice-cert major:0 minor:1039 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/2c3fab4d-05b8-42ec-8c91-91ff64562649/volumes/kubernetes.io~secret/webhook-cert:{mountpoint:/var/lib/kubelet/pods/2c3fab4d-05b8-42ec-8c91-91ff64562649/volumes/kubernetes.io~secret/webhook-cert major:0 minor:1035 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/3332b604-6f0d-4243-bce1-b3ab76e58c78/volumes/kubernetes.io~projected/kube-api-access-6gfgm:{mountpoint:/var/lib/kubelet/pods/3332b604-6f0d-4243-bce1-b3ab76e58c78/volumes/kubernetes.io~projected/kube-api-access-6gfgm major:0 minor:1637 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/3332b604-6f0d-4243-bce1-b3ab76e58c78/volumes/kubernetes.io~secret/webhook-certs:{mountpoint:/var/lib/kubelet/pods/3332b604-6f0d-4243-bce1-b3ab76e58c78/volumes/kubernetes.io~secret/webhook-certs major:0 minor:1636 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/33616deb-ea10-4a38-8681-ab023b526b11/volumes/kubernetes.io~empty-dir/etc-tuned:{mountpoint:/var/lib/kubelet/pods/33616deb-ea10-4a38-8681-ab023b526b11/volumes/kubernetes.io~empty-dir/etc-tuned major:0 minor:581 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/33616deb-ea10-4a38-8681-ab023b526b11/volumes/kubernetes.io~empty-dir/tmp:{mountpoint:/var/lib/kubelet/pods/33616deb-ea10-4a38-8681-ab023b526b11/volumes/kubernetes.io~empty-dir/tmp major:0 minor:580 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/33616deb-ea10-4a38-8681-ab023b526b11/volumes/kubernetes.io~projected/kube-api-access-rsrkp:{mountpoint:/var/lib/kubelet/pods/33616deb-ea10-4a38-8681-ab023b526b11/volumes/kubernetes.io~projected/kube-api-access-rsrkp major:0 minor:582 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~projected/kube-api-access-lgtnt:{mountpoint:/var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~projected/kube-api-access-lgtnt major:0 minor:1022 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/federate-client-tls:{mountpoint:/var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/federate-client-tls major:0 minor:782 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/secret-telemeter-client:{mountpoint:/var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/secret-telemeter-client major:0 minor:785 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/secret-telemeter-client-kube-rbac-proxy-config:{mountpoint:/var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/secret-telemeter-client-kube-rbac-proxy-config major:0 minor:784 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/telemeter-client-tls:{mountpoint:/var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/telemeter-client-tls major:0 minor:783 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8/volumes/kubernetes.io~projected/kube-api-access-jrdsv:{mountpoint:/var/lib/kubelet/pods/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8/volumes/kubernetes.io~projected/kube-api-access-jrdsv major:0 minor:138 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c/volumes/kubernetes.io~projected/kube-api-access-ssssf:{mountpoint:/var/lib/kubelet/pods/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c/volumes/kubernetes.io~projected/kube-api-access-ssssf major:0 minor:153 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c/volumes/kubernetes.io~secret/metrics-certs:{mountpoint:/var/lib/kubelet/pods/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c/volumes/kubernetes.io~secret/metrics-certs major:0 minor:614 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~projected/kube-api-access-5rlts:{mountpoint:/var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~projected/kube-api-access-5rlts major:0 minor:1559 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~secret/client-ca-bundle:{mountpoint:/var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~secret/client-ca-bundle major:0 minor:1558 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~secret/secret-metrics-client-certs:{mountpoint:/var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~secret/secret-metrics-client-certs major:0 minor:1557 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~secret/secret-metrics-server-tls:{mountpoint:/var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~secret/secret-metrics-server-tls major:0 minor:1556 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a/volumes/kubernetes.io~projected/kube-api-access-pdsh9:{mountpoint:/var/lib/kubelet/pods/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a/volumes/kubernetes.io~projected/kube-api-access-pdsh9 major:0 minor:472 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~projected/kube-api-access:{mountpoint:/var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~projected/kube-api-access major:0 minor:288 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~secret/serving-cert major:0 minor:260 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~projected/kube-api-access-hskm2:{mountpoint:/var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~projected/kube-api-access-hskm2 major:0 minor:1341 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~secret/default-certificate:{mountpoint:/var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~secret/default-certificate major:0 minor:1335 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~secret/metrics-certs:{mountpoint:/var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~secret/metrics-certs major:0 minor:1336 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~secret/stats-auth:{mountpoint:/var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~secret/stats-auth major:0 minor:1338 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~projected/kube-api-access-jv4zs:{mountpoint:/var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~projected/kube-api-access-jv4zs major:0 minor:286 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~secret/profile-collector-cert:{mountpoint:/var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~secret/profile-collector-cert major:0 minor:254 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~secret/srv-cert:{mountpoint:/var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~secret/srv-cert major:0 minor:616 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~projected/kube-api-access:{mountpoint:/var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~projected/kube-api-access major:0 minor:290 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~secret/serving-cert major:0 minor:258 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/489b6771-aa16-469f-9883-9a9233d3c379/volumes/kubernetes.io~projected/kube-api-access-9pvrn:{mountpoint:/var/lib/kubelet/pods/489b6771-aa16-469f-9883-9a9233d3c379/volumes/kubernetes.io~projected/kube-api-access-9pvrn major:0 minor:1166 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/48bd1d86-a6f2-439f-ab04-6a9a442bec42/volumes/kubernetes.io~projected/kube-api-access-f69hj:{mountpoint:/var/lib/kubelet/pods/48bd1d86-a6f2-439f-ab04-6a9a442bec42/volumes/kubernetes.io~projected/kube-api-access-f69hj major:0 minor:1045 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/48bd1d86-a6f2-439f-ab04-6a9a442bec42/volumes/kubernetes.io~secret/cert:{mountpoint:/var/lib/kubelet/pods/48bd1d86-a6f2-439f-ab04-6a9a442bec42/volumes/kubernetes.io~secret/cert major:0 minor:1036 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/48bd1d86-a6f2-439f-ab04-6a9a442bec42/volumes/kubernetes.io~secret/cluster-baremetal-operator-tls:{mountpoint:/var/lib/kubelet/pods/48bd1d86-a6f2-439f-ab04-6a9a442bec42/volumes/kubernetes.io~secret/cluster-baremetal-operator-tls major:0 minor:1041 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~projected/kube-api-access-6vp9b:{mountpoint:/var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~projected/kube-api-access-6vp9b major:0 minor:289 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~secret/cluster-olm-operator-serving-cert:{mountpoint:/var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~secret/cluster-olm-operator-serving-cert major:0 minor:263 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volume-subpaths/run-systemd/ovnkube-controller/6:{mountpoint:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volume-subpaths/run-systemd/ovnkube-controller/6 major:0 minor:24 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~projected/kube-api-access-tjgc4:{mountpoint:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~projected/kube-api-access-tjgc4 major:0 minor:172 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~secret/ovn-node-metrics-cert:{mountpoint:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~secret/ovn-node-metrics-cert major:0 minor:171 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/4c2975ec-e33d-4960-a708-277d41c79b15/volumes/kubernetes.io~projected/kube-api-access-d6z8n:{mountpoint:/var/lib/kubelet/pods/4c2975ec-e33d-4960-a708-277d41c79b15/volumes/kubernetes.io~projected/kube-api-access-d6z8n major:0 minor:1046 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/4c2975ec-e33d-4960-a708-277d41c79b15/volumes/kubernetes.io~secret/profile-collector-cert:{mountpoint:/var/lib/kubelet/pods/4c2975ec-e33d-4960-a708-277d41c79b15/volumes/kubernetes.io~secret/profile-collector-cert major:0 minor:1031 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/4c2975ec-e33d-4960-a708-277d41c79b15/volumes/kubernetes.io~secret/srv-cert:{mountpoint:/var/lib/kubelet/pods/4c2975ec-e33d-4960-a708-277d41c79b15/volumes/kubernetes.io~secret/srv-cert major:0 minor:1034 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/54f9f9d4-8c31-4636-8010-22ee11b9b323/volumes/kubernetes.io~projected/kube-api-access-sv2tg:{mountpoint:/var/lib/kubelet/pods/54f9f9d4-8c31-4636-8010-22ee11b9b323/volumes/kubernetes.io~projected/kube-api-access-sv2tg major:0 minor:1475 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/54f9f9d4-8c31-4636-8010-22ee11b9b323/volumes/kubernetes.io~secret/kube-state-metrics-kube-rbac-proxy-config:{mountpoint:/var/lib/kubelet/pods/54f9f9d4-8c31-4636-8010-22ee11b9b323/volumes/kubernetes.io~secret/kube-state-metrics-kube-rbac-proxy-config major:0 minor:1473 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/54f9f9d4-8c31-4636-8010-22ee11b9b323/volumes/kubernetes.io~secret/kube-state-metrics-tls:{mountpoint:/var/lib/kubelet/pods/54f9f9d4-8c31-4636-8010-22ee11b9b323/volumes/kubernetes.io~secret/kube-state-metrics-tls major:0 minor:1472 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/568e9bed-cbd2-49bd-84af-9e17f95a003f/volumes/kubernetes.io~projected/kube-api-access-flvlq:{mountpoint:/var/lib/kubelet/pods/568e9bed-cbd2-49bd-84af-9e17f95a003f/volumes/kubernetes.io~projected/kube-api-access-flvlq major:0 minor:1364 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/568e9bed-cbd2-49bd-84af-9e17f95a003f/volumes/kubernetes.io~secret/certs:{mountpoint:/var/lib/kubelet/pods/568e9bed-cbd2-49bd-84af-9e17f95a003f/volumes/kubernetes.io~secret/certs major:0 minor:1362 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/568e9bed-cbd2-49bd-84af-9e17f95a003f/volumes/kubernetes.io~secret/node-bootstrap-token:{mountpoint:/var/lib/kubelet/pods/568e9bed-cbd2-49bd-84af-9e17f95a003f/volumes/kubernetes.io~secret/node-bootstrap-token major:0 minor:1363 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/58b95ae4-7f74-4777-8441-0a0ae28199e9/volumes/kubernetes.io~projected/kube-api-access-7fbtj:{mountpoint:/var/lib/kubelet/pods/58b95ae4-7f74-4777-8441-0a0ae28199e9/volumes/kubernetes.io~projected/kube-api-access-7fbtj major:0 minor:1047 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/58b95ae4-7f74-4777-8441-0a0ae28199e9/volumes/kubernetes.io~secret/cluster-storage-operator-serving-cert:{mountpoint:/var/lib/kubelet/pods/58b95ae4-7f74-4777-8441-0a0ae28199e9/volumes/kubernetes.io~secret/cluster-storage-operator-serving-cert major:0 minor:1033 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/58baad85-de54-49e7-a13e-d470d9c50d11/volumes/kubernetes.io~projected/kube-api-access-jx7jr:{mountpoint:/var/lib/kubelet/pods/58baad85-de54-49e7-a13e-d470d9c50d11/volumes/kubernetes.io~projected/kube-api-access-jx7jr major:0 minor:722 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/58baad85-de54-49e7-a13e-d470d9c50d11/volumes/kubernetes.io~secret/metrics-tls:{mountpoint:/var/lib/kubelet/pods/58baad85-de54-49e7-a13e-d470d9c50d11/volumes/kubernetes.io~secret/metrics-tls major:0 minor:739 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/592373ae-a952-4dd3-a9bc-f9c9c19c0802/volumes/kubernetes.io~projected/ca-certs:{mountpoint:/var/lib/kubelet/pods/592373ae-a952-4dd3-a9bc-f9c9c19c0802/volumes/kubernetes.io~projected/ca-certs major:0 minor:819 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/592373ae-a952-4dd3-a9bc-f9c9c19c0802/volumes/kubernetes.io~projected/kube-api-access-rd87p:{mountpoint:/var/lib/kubelet/pods/592373ae-a952-4dd3-a9bc-f9c9c19c0802/volumes/kubernetes.io~projected/kube-api-access-rd87p major:0 minor:818 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/5a1bdc70-6412-47e0-8330-04d796cc8d55/volumes/kubernetes.io~projected/kube-api-access-5dtfn:{mountpoint:/var/lib/kubelet/pods/5a1bdc70-6412-47e0-8330-04d796cc8d55/volumes/kubernetes.io~projected/kube-api-access-5dtfn major:0 minor:292 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/5a1bdc70-6412-47e0-8330-04d796cc8d55/volumes/kubernetes.io~secret/metrics-tls:{mountpoint:/var/lib/kubelet/pods/5a1bdc70-6412-47e0-8330-04d796cc8d55/volumes/kubernetes.io~secret/metrics-tls major:0 minor:557 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/62f174ff-9f3c-4051-a34a-e55a59207171/volumes/kubernetes.io~projected/kube-api-access-r5p8s:{mountpoint:/var/lib/kubelet/pods/62f174ff-9f3c-4051-a34a-e55a59207171/volumes/kubernetes.io~projected/kube-api-access-r5p8s major:0 minor:1145 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/62f174ff-9f3c-4051-a34a-e55a59207171/volumes/kubernetes.io~secret/machine-approver-tls:{mountpoint:/var/lib/kubelet/pods/62f174ff-9f3c-4051-a34a-e55a59207171/volumes/kubernetes.io~secret/machine-approver-tls major:0 minor:1144 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd/volumes/kubernetes.io~projected/kube-api-access-vtsw9:{mountpoint:/var/lib/kubelet/pods/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd/volumes/kubernetes.io~projected/kube-api-access-vtsw9 major:0 minor:1054 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd/volumes/kubernetes.io~secret/serving-cert major:0 minor:1038 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/6728a458-098b-44f9-8c8f-b0d76c5825fa/volumes/kubernetes.io~projected/kube-api-access-784mb:{mountpoint:/var/lib/kubelet/pods/6728a458-098b-44f9-8c8f-b0d76c5825fa/volumes/kubernetes.io~projected/kube-api-access-784mb major:0 minor:1311 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/6728a458-098b-44f9-8c8f-b0d76c5825fa/volumes/kubernetes.io~secret/proxy-tls:{mountpoint:/var/lib/kubelet/pods/6728a458-098b-44f9-8c8f-b0d76c5825fa/volumes/kubernetes.io~secret/proxy-tls major:0 minor:1310 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/6a9f011f-36f1-4308-a365-69425c186c7f/volumes/kubernetes.io~projected/kube-api-access-mvnxf:{mountpoint:/var/lib/kubelet/pods/6a9f011f-36f1-4308-a365-69425c186c7f/volumes/kubernetes.io~projected/kube-api-access-mvnxf major:0 minor:137 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes.io~projected/kube-api-access-fzwfq:{mountpoint:/var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes.io~projected/kube-api-access-fzwfq major:0 minor:266 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes Dec 05 10:47:20.072709 master-0 kubenswrapper[24928]: .io~secret/serving-cert major:0 minor:264 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/720a7467-ce93-4d48-82ec-9ad0922d99c2/volumes/kubernetes.io~projected/kube-api-access-5n7tf:{mountpoint:/var/lib/kubelet/pods/720a7467-ce93-4d48-82ec-9ad0922d99c2/volumes/kubernetes.io~projected/kube-api-access-5n7tf major:0 minor:482 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/82464e92-4de2-48f0-8772-a489abb16898/volumes/kubernetes.io~projected/kube-api-access-4dk2b:{mountpoint:/var/lib/kubelet/pods/82464e92-4de2-48f0-8772-a489abb16898/volumes/kubernetes.io~projected/kube-api-access-4dk2b major:0 minor:1052 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/82464e92-4de2-48f0-8772-a489abb16898/volumes/kubernetes.io~secret/machine-api-operator-tls:{mountpoint:/var/lib/kubelet/pods/82464e92-4de2-48f0-8772-a489abb16898/volumes/kubernetes.io~secret/machine-api-operator-tls major:0 minor:1032 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/87909f47-f2d7-46f8-a1c8-27336cdcce5d/volumes/kubernetes.io~projected/kube-api-access-rjrgm:{mountpoint:/var/lib/kubelet/pods/87909f47-f2d7-46f8-a1c8-27336cdcce5d/volumes/kubernetes.io~projected/kube-api-access-rjrgm major:0 minor:323 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/88cccb5b-1ad1-4fab-b34c-90252794ee20/volumes/kubernetes.io~projected/kube-api-access-f4lbg:{mountpoint:/var/lib/kubelet/pods/88cccb5b-1ad1-4fab-b34c-90252794ee20/volumes/kubernetes.io~projected/kube-api-access-f4lbg major:0 minor:953 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/88cccb5b-1ad1-4fab-b34c-90252794ee20/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/88cccb5b-1ad1-4fab-b34c-90252794ee20/volumes/kubernetes.io~secret/serving-cert major:0 minor:952 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/8c649a16-c187-412e-b5da-62a00bee38ab/volumes/kubernetes.io~projected/kube-api-access-d8pv2:{mountpoint:/var/lib/kubelet/pods/8c649a16-c187-412e-b5da-62a00bee38ab/volumes/kubernetes.io~projected/kube-api-access-d8pv2 major:0 minor:309 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/8c649a16-c187-412e-b5da-62a00bee38ab/volumes/kubernetes.io~secret/package-server-manager-serving-cert:{mountpoint:/var/lib/kubelet/pods/8c649a16-c187-412e-b5da-62a00bee38ab/volumes/kubernetes.io~secret/package-server-manager-serving-cert major:0 minor:613 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/926263c4-ec5b-41cb-9c30-0c88f636035f/volumes/kubernetes.io~projected/kube-api-access-fncz7:{mountpoint:/var/lib/kubelet/pods/926263c4-ec5b-41cb-9c30-0c88f636035f/volumes/kubernetes.io~projected/kube-api-access-fncz7 major:0 minor:318 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/926263c4-ec5b-41cb-9c30-0c88f636035f/volumes/kubernetes.io~secret/marketplace-operator-metrics:{mountpoint:/var/lib/kubelet/pods/926263c4-ec5b-41cb-9c30-0c88f636035f/volumes/kubernetes.io~secret/marketplace-operator-metrics major:0 minor:615 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/97192d4d-15d3-4740-82f3-d0d45f9fe7b9/volumes/kubernetes.io~projected/kube-api-access-nn6sc:{mountpoint:/var/lib/kubelet/pods/97192d4d-15d3-4740-82f3-d0d45f9fe7b9/volumes/kubernetes.io~projected/kube-api-access-nn6sc major:0 minor:500 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/97192d4d-15d3-4740-82f3-d0d45f9fe7b9/volumes/kubernetes.io~secret/signing-key:{mountpoint:/var/lib/kubelet/pods/97192d4d-15d3-4740-82f3-d0d45f9fe7b9/volumes/kubernetes.io~secret/signing-key major:0 minor:498 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/a4fd453c-e667-4bdc-aa9e-3d95ff707200/volumes/kubernetes.io~projected/ca-certs:{mountpoint:/var/lib/kubelet/pods/a4fd453c-e667-4bdc-aa9e-3d95ff707200/volumes/kubernetes.io~projected/ca-certs major:0 minor:816 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/a4fd453c-e667-4bdc-aa9e-3d95ff707200/volumes/kubernetes.io~projected/kube-api-access-kmhhw:{mountpoint:/var/lib/kubelet/pods/a4fd453c-e667-4bdc-aa9e-3d95ff707200/volumes/kubernetes.io~projected/kube-api-access-kmhhw major:0 minor:817 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/a4fd453c-e667-4bdc-aa9e-3d95ff707200/volumes/kubernetes.io~secret/catalogserver-certs:{mountpoint:/var/lib/kubelet/pods/a4fd453c-e667-4bdc-aa9e-3d95ff707200/volumes/kubernetes.io~secret/catalogserver-certs major:0 minor:815 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/a722cda9-29a0-4b7f-8e1d-9a8950ed765a/volumes/kubernetes.io~projected/kube-api-access-2lwgq:{mountpoint:/var/lib/kubelet/pods/a722cda9-29a0-4b7f-8e1d-9a8950ed765a/volumes/kubernetes.io~projected/kube-api-access-2lwgq major:0 minor:276 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/a722cda9-29a0-4b7f-8e1d-9a8950ed765a/volumes/kubernetes.io~secret/cluster-monitoring-operator-tls:{mountpoint:/var/lib/kubelet/pods/a722cda9-29a0-4b7f-8e1d-9a8950ed765a/volumes/kubernetes.io~secret/cluster-monitoring-operator-tls major:0 minor:617 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/af7078c2-ad4f-415b-ba2f-77f5145c3b3f/volumes/kubernetes.io~projected/kube-api-access-ljdt5:{mountpoint:/var/lib/kubelet/pods/af7078c2-ad4f-415b-ba2f-77f5145c3b3f/volumes/kubernetes.io~projected/kube-api-access-ljdt5 major:0 minor:1055 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/af7078c2-ad4f-415b-ba2f-77f5145c3b3f/volumes/kubernetes.io~secret/cloud-controller-manager-operator-tls:{mountpoint:/var/lib/kubelet/pods/af7078c2-ad4f-415b-ba2f-77f5145c3b3f/volumes/kubernetes.io~secret/cloud-controller-manager-operator-tls major:0 minor:1040 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e/volumes/kubernetes.io~projected/kube-api-access-g88bl:{mountpoint:/var/lib/kubelet/pods/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e/volumes/kubernetes.io~projected/kube-api-access-g88bl major:0 minor:1444 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e/volumes/kubernetes.io~secret/prometheus-operator-kube-rbac-proxy-config:{mountpoint:/var/lib/kubelet/pods/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e/volumes/kubernetes.io~secret/prometheus-operator-kube-rbac-proxy-config major:0 minor:1443 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e/volumes/kubernetes.io~secret/prometheus-operator-tls:{mountpoint:/var/lib/kubelet/pods/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e/volumes/kubernetes.io~secret/prometheus-operator-tls major:0 minor:1445 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/b7508cd3-7421-4ab9-be1f-318db2853596/volumes/kubernetes.io~projected/kube-api-access-kgvv6:{mountpoint:/var/lib/kubelet/pods/b7508cd3-7421-4ab9-be1f-318db2853596/volumes/kubernetes.io~projected/kube-api-access-kgvv6 major:0 minor:644 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/b7508cd3-7421-4ab9-be1f-318db2853596/volumes/kubernetes.io~secret/cert:{mountpoint:/var/lib/kubelet/pods/b7508cd3-7421-4ab9-be1f-318db2853596/volumes/kubernetes.io~secret/cert major:0 minor:645 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~projected/kube-api-access-jfz87:{mountpoint:/var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~projected/kube-api-access-jfz87 major:0 minor:164 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~secret/ovn-control-plane-metrics-cert:{mountpoint:/var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~secret/ovn-control-plane-metrics-cert major:0 minor:163 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/b7e7f216-f9af-41df-a0b0-df2a76b9b72a/volumes/kubernetes.io~projected/kube-api-access-rvw27:{mountpoint:/var/lib/kubelet/pods/b7e7f216-f9af-41df-a0b0-df2a76b9b72a/volumes/kubernetes.io~projected/kube-api-access-rvw27 major:0 minor:1170 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/b7e7f216-f9af-41df-a0b0-df2a76b9b72a/volumes/kubernetes.io~secret/proxy-tls:{mountpoint:/var/lib/kubelet/pods/b7e7f216-f9af-41df-a0b0-df2a76b9b72a/volumes/kubernetes.io~secret/proxy-tls major:0 minor:1169 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/baa812e4-b011-41ee-82ab-8f571f0b7e0a/volumes/kubernetes.io~projected/kube-api-access-28qvt:{mountpoint:/var/lib/kubelet/pods/baa812e4-b011-41ee-82ab-8f571f0b7e0a/volumes/kubernetes.io~projected/kube-api-access-28qvt major:0 minor:1167 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/baee05cb-62ad-4eda-8a13-e317eae840e5/volumes/kubernetes.io~projected/kube-api-access-564nt:{mountpoint:/var/lib/kubelet/pods/baee05cb-62ad-4eda-8a13-e317eae840e5/volumes/kubernetes.io~projected/kube-api-access-564nt major:0 minor:726 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~projected/kube-api-access-98qt8:{mountpoint:/var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~projected/kube-api-access-98qt8 major:0 minor:273 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~secret/serving-cert major:0 minor:259 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~projected/kube-api-access-ltjt6:{mountpoint:/var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~projected/kube-api-access-ltjt6 major:0 minor:811 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~secret/encryption-config:{mountpoint:/var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~secret/encryption-config major:0 minor:809 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~secret/etcd-client:{mountpoint:/var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~secret/etcd-client major:0 minor:808 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~secret/serving-cert major:0 minor:810 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~projected/kube-api-access-fkmvj:{mountpoint:/var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~projected/kube-api-access-fkmvj major:0 minor:274 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~secret/serving-cert major:0 minor:265 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~projected/kube-api-access-m88gw:{mountpoint:/var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~projected/kube-api-access-m88gw major:0 minor:611 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~secret/encryption-config:{mountpoint:/var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~secret/encryption-config major:0 minor:561 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~secret/etcd-client:{mountpoint:/var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~secret/etcd-client major:0 minor:562 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~secret/serving-cert major:0 minor:733 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~projected/kube-api-access-wg9zq:{mountpoint:/var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~projected/kube-api-access-wg9zq major:0 minor:188 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~secret/webhook-cert:{mountpoint:/var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~secret/webhook-cert major:0 minor:187 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~projected/kube-api-access-2fbss:{mountpoint:/var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~projected/kube-api-access-2fbss major:0 minor:313 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~secret/serving-cert major:0 minor:262 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d9ff1ae2-8547-42db-9fd0-f4782589ca18/volumes/kubernetes.io~projected/kube-api-access-wpgqz:{mountpoint:/var/lib/kubelet/pods/d9ff1ae2-8547-42db-9fd0-f4782589ca18/volumes/kubernetes.io~projected/kube-api-access-wpgqz major:0 minor:1469 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d9ff1ae2-8547-42db-9fd0-f4782589ca18/volumes/kubernetes.io~secret/openshift-state-metrics-kube-rbac-proxy-config:{mountpoint:/var/lib/kubelet/pods/d9ff1ae2-8547-42db-9fd0-f4782589ca18/volumes/kubernetes.io~secret/openshift-state-metrics-kube-rbac-proxy-config major:0 minor:1468 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/d9ff1ae2-8547-42db-9fd0-f4782589ca18/volumes/kubernetes.io~secret/openshift-state-metrics-tls:{mountpoint:/var/lib/kubelet/pods/d9ff1ae2-8547-42db-9fd0-f4782589ca18/volumes/kubernetes.io~secret/openshift-state-metrics-tls major:0 minor:1467 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/da9d48c9-6346-4c9f-a690-f7419499c3e6/volumes/kubernetes.io~projected/kube-api-access-7824l:{mountpoint:/var/lib/kubelet/pods/da9d48c9-6346-4c9f-a690-f7419499c3e6/volumes/kubernetes.io~projected/kube-api-access-7824l major:0 minor:1339 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0/volumes/kubernetes.io~projected/kube-api-access-95jmp:{mountpoint:/var/lib/kubelet/pods/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0/volumes/kubernetes.io~projected/kube-api-access-95jmp major:0 minor:860 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0/volumes/kubernetes.io~secret/serving-cert major:0 minor:832 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/volumes/kubernetes.io~projected/kube-api-access-ml74w:{mountpoint:/var/lib/kubelet/pods/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/volumes/kubernetes.io~projected/kube-api-access-ml74w major:0 minor:465 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~projected/kube-api-access-qn7rj:{mountpoint:/var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~projected/kube-api-access-qn7rj major:0 minor:267 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~secret/apiservice-cert:{mountpoint:/var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~secret/apiservice-cert major:0 minor:524 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~secret/node-tuning-operator-tls:{mountpoint:/var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~secret/node-tuning-operator-tls major:0 minor:523 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/ee0b4a7f-1b96-4304-bef0-fb575e0e431c/volumes/kubernetes.io~projected/kube-api-access-9ktnb:{mountpoint:/var/lib/kubelet/pods/ee0b4a7f-1b96-4304-bef0-fb575e0e431c/volumes/kubernetes.io~projected/kube-api-access-9ktnb major:0 minor:1053 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/ee0b4a7f-1b96-4304-bef0-fb575e0e431c/volumes/kubernetes.io~secret/cloud-credential-operator-serving-cert:{mountpoint:/var/lib/kubelet/pods/ee0b4a7f-1b96-4304-bef0-fb575e0e431c/volumes/kubernetes.io~secret/cloud-credential-operator-serving-cert major:0 minor:1042 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~projected/kube-api-access-twh85:{mountpoint:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~projected/kube-api-access-twh85 major:0 minor:299 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/etcd-client:{mountpoint:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/etcd-client major:0 minor:261 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/serving-cert major:0 minor:255 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/bound-sa-token:{mountpoint:/var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/bound-sa-token major:0 minor:312 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/kube-api-access-khfxz:{mountpoint:/var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/kube-api-access-khfxz major:0 minor:296 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~secret/image-registry-operator-tls:{mountpoint:/var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~secret/image-registry-operator-tls major:0 minor:558 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~projected/kube-api-access-h2vd4:{mountpoint:/var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~projected/kube-api-access-h2vd4 major:0 minor:325 fsType:tmpfs blockSize:0} /var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~secret/serving-cert:{mountpoint:/var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~secret/serving-cert major:0 minor:277 fsType:tmpfs blockSize:0} overlay_0-100:{mountpoint:/var/lib/containers/storage/overlay/37f7404a5f952c26e4c10d3bb5e10c38f796e2f60279aa4b358fa3499c9dc3ce/merged major:0 minor:100 fsType:overlay blockSize:0} overlay_0-1006:{mountpoint:/var/lib/containers/storage/overlay/e7d0dcdb5721b512dd2a3811a93b2d429834e4cf4a9ddc0fed97c8485bbf9717/merged major:0 minor:1006 fsType:overlay blockSize:0} overlay_0-1008:{mountpoint:/var/lib/containers/storage/overlay/356010bebd060051f4cc748afe3f99dc01d2ff8de18331ad388ac44e233d1304/merged major:0 minor:1008 fsType:overlay blockSize:0} overlay_0-1011:{mountpoint:/var/lib/containers/storage/overlay/ffb894f08f387d1b69c6409713d3369aa258e2fbd1114c44b3c9629a372f1e16/merged major:0 minor:1011 fsType:overlay blockSize:0} overlay_0-102:{mountpoint:/var/lib/containers/storage/overlay/9741f6d1f71bfd0fc387b1bbea32f99c3d5c3f8a4f50ebbde5e06e3db8e60590/merged major:0 minor:102 fsType:overlay blockSize:0} overlay_0-1020:{mountpoint:/var/lib/containers/storage/overlay/92fe9a915a9dcedea8c899ac6e9a125c41716e08bafa68a20a516598608267be/merged major:0 minor:1020 fsType:overlay blockSize:0} overlay_0-1025:{mountpoint:/var/lib/containers/storage/overlay/2fdae9e4459d2e0e88db303039c6b180520ecc0f0c987a016d46fdcd138203a9/merged major:0 minor:1025 fsType:overlay blockSize:0} overlay_0-104:{mountpoint:/var/lib/containers/storage/overlay/cc1260dd9574751486fb09f742d5c780075da35fdada0b9b649da084f419e84d/merged major:0 minor:104 fsType:overlay blockSize:0} overlay_0-1058:{mountpoint:/var/lib/containers/storage/overlay/783e9dc7983e85bc31b89f63274df8f1e2bd2e542ece36b7ce9938bee3bd27c4/merged major:0 minor:1058 fsType:overlay blockSize:0} overlay_0-106:{mountpoint:/var/lib/containers/storage/overlay/7b11865fedac6857639cb41a47ddf885d1aec695a812559ba44daabf9e833596/merged major:0 minor:106 fsType:overlay blockSize:0} overlay_0-108:{mountpoint:/var/lib/containers/storage/overlay/c2eef8519e1a600a7c959207dc523df29ded81b900458f0d14ec4518afdb5454/merged major:0 minor:108 fsType:overlay blockSize:0} overlay_0-1080:{mountpoint:/var/lib/containers/storage/overlay/e816eea4fedc03fc3e793ad8ddccaba5260e3b7828af8edc1d86517dda74ea90/merged major:0 minor:1080 fsType:overlay blockSize:0} overlay_0-1082:{mountpoint:/var/lib/containers/storage/overlay/08241b02119eb93448d2e7bb949a9f4527501e8f1ef418ba773b0323731c2ad5/merged major:0 minor:1082 fsType:overlay blockSize:0} overlay_0-1095:{mountpoint:/var/lib/containers/storage/overlay/1f641f11d19670d4b29b30a09ae326c5ae6b51461cdf0c0780a57cf4110a892c/merged major:0 minor:1095 fsType:overlay blockSize:0} overlay_0-1097:{mountpoint:/var/lib/containers/storage/overlay/6083fe16e69228bc524d1d4eb606255d97983baabdbcbbe3bda8b482c194a3a0/merged major:0 minor:1097 fsType:overlay blockSize:0} overlay_0-1099:{mountpoint:/var/lib/containers/storage/overlay/79e486003443d9492c2b556244c5ae98550f536051047a6e60d7d487e0cd5578/merged major:0 minor:1099 fsType:overlay blockSize:0} overlay_0-110:{mountpoint:/var/lib/containers/storage/overlay/784d59f45a1260cfbad712ec9faa218b405c620f7aa0a00a602b8845a66329f2/merged major:0 minor:110 fsType:overlay blockSize:0} overlay_0-1101:{mountpoint:/var/lib/containers/storage/overlay/302054f65733db16d6df62fd993b3d6281146d083e78a835e009369de24f5b8d/merged major:0 minor:1101 fsType:overlay blockSize:0} overlay_0-1103:{mountpoint:/var/lib/containers/storage/overlay/6b4e4114f652164a9637fc042b4167ef57075c3dcf2c0e42ee734ebd963bec6b/merged major:0 minor:1103 fsType:overlay blockSize:0} overlay_0-1105:{mountpoint:/var/lib/containers/storage/overlay/a8ea2c592e9ceea6d89a568f0ba0bc9e5e02811f324283bc48d8c33b1cd3a4b1/merged major:0 minor:1105 fsType:overlay blockSize:0} overlay_0-1119:{mountpoint:/var/lib/containers/storage/overlay/d388d6234f147797bd1edd4e57415ef0da902193e0361ceb75793f062df117e8/merged major:0 minor:1119 fsType:overlay blockSize:0} overlay_0-112:{mountpoint:/var/lib/containers/storage/overlay/e291e4109097f244830882c6e9a1bee5529a9fddf32e01b243b7273d78e1bc4e/merged major:0 minor:112 fsType:overlay blockSize:0} overlay_0-1121:{mountpoint:/var/lib/containers/storage/overlay/1a4d2c9498893842072d78a6f7f9c51956fc56c0ad256ba9c64e4566158d713b/merged major:0 minor:1121 fsType:overlay blockSize:0} overlay_0-1123:{mountpoint:/var/lib/containers/storage/overlay/3b40dd4e4c473fddb3cf66760f1588d5653e69ede8f3237f907e1f592f101dfe/merged major:0 minor:1123 fsType:overlay blockSize:0} overlay_0-1125:{mountpoint:/var/lib/containers/storage/overlay/6afddee5c70a59a1e3e0946b3f07c2a57c9da46215bd2ec28afb03efaa352ec1/merged major:0 minor:1125 fsType:overlay blockSize:0} overlay_0-1148:{mountpoint:/var/lib/containers/storage/overlay/a0f67b807ef4512546450be86e822f6293b5e8c5b463c6513892abd99703fd6a/merged major:0 minor:1148 fsType:overlay blockSize:0} overlay_0-1150:{mountpoint:/var/lib/containers/storage/overlay/61c7cc608d5bc7d0ae56137c8b568cafed132b4c47089699fc1853ca9995b83d/merged major:0 minor:1150 fsType:overlay blockSize:0} overlay_0-1158:{mountpoint:/var/lib/containers/storage/overlay/2e3ec8f6282db1527b450328d9aa2dce9bc4c1245da8d5c5a667e533c4e6ad41/merged major:0 minor:1158 fsType:overlay blockSize:0} overlay_0-1181:{mountpoint:/var/lib/containers/storage/overlay/c8fc4811789a88945f6609a08b431887865e927047bc46c27f9b5e23e18bd3ba/merged major:0 minor:1181 fsType:overlay blockSize:0} overlay_0-1184:{mountpoint:/var/lib/containers/storage/overlay/547ea691aaaf186aa155183fb92e81460bebe82d9e9ca9e357df29f69af496d0/merged major:0 minor:1184 fsType:overlay blockSize:0} overlay_0-1186:{mountpoint:/var/lib/containers/storage/overlay/5d8d123c23ce77017fa01c65743ca40221830274c8cd15b73c471a64b6afe660/merged major:0 minor:1186 fsType:overlay blockSize:0} overlay_0-1188:{mountpoint:/var/lib/containers/storage/overlay/bd4c3fc039bcdf771f447e3efa72f69a4f7c029e7b12510dc5a0f7635dd55e34/merged major:0 minor:1188 fsType:overlay blockSize:0} overlay_0-1190:{mountpoint:/var/lib/containers/storage/overlay/35d599139be46488029bee682cb8f43ce89901d6662e4955c005ce02025ac488/merged major:0 minor:1190 fsType:overlay blockSize:0} overlay_0-1192:{mountpoint:/var/lib/containers/storage/overlay/9f681051e63581e4759b4a5a8efa086454dcec58fcea5cb8111a73c903a7bbc6/merged major:0 minor:1192 fsType:overlay blockSize:0} overlay_0-1195:{mountpoint:/var/lib/containers/storage/overlay/04e0a92bdddf2e98bab8fb1abe8674c0af0daf28aca06cc7bcb5186f266db17e/merged major:0 minor:1195 fsType:overlay blockSize:0} overlay_0-120:{mountpoint:/var/lib/containers/storage/overlay/d1e62255e12c370b98430fede60367d62bee912ad46005c4d932d9c04e44d755/merged major:0 minor:120 fsType:overlay blockSize:0} overlay_0-1202:{mountpoint:/var/lib/containers/storage/overlay/8035e30cdc5a7412d2dcd29351b4ee81c7bd5f51cc8179026ac8f6f82fd0145d/merged major:0 minor:1202 fsType:overlay blockSize:0} overlay_0-1204:{mountpoint:/var/lib/containers/storage/overlay/eed0d47b04de1c64aac79c778ab451df9d98243b69ab344fce0dae35a8be3bda/merged major:0 minor:1204 fsType:overlay blockSize:0} overlay_0-1206:{mountpoint:/var/lib/containers/storage/overlay/b8435e79f39544d22fb884fa9ddfcfea7eb2f18b5e096693c8f09dc6dcba8dd7/merged major:0 minor:1206 fsType:overlay blockSize:0} overlay_0-1208:{mountpoint:/var/lib/containers/storage/overlay/c6156916451d101b813c45225e006c961e5d3f631afa89e330dd7b9ca4603a98/merged major:0 minor:1208 fsType:overlay blockSize:0} overlay_0-1264:{mountpoint:/var/lib/containers/storage/overlay/cc1685d0e6770608d7b80fcb37d05b78f62493178571622028853e46d03d7511/merged major:0 minor:1264 fsType:overlay blockSize:0} overlay_0-1266:{mountpoint:/var/lib/containers/storage/overlay/5e50d0eabddb5a69cfdf4689b088ad826a2027dd7cf2eeacc8e5aea20d780b99/merged major:0 minor:1266 fsType:overlay blockSize:0} overlay_0-1268:{mountpoint:/var/lib/containers/storage/overlay/4e23b52ab992458229602a212de5258fc62447beb17e34199eff532d2809f374/merged major:0 minor:1268 fsType:overlay blockSize:0} overlay_0-1270:{mountpoint:/var/lib/containers/storage/overlay/231408eb2fb884a545fc1e847b35774ebcdc227b42bb19f10ce0c84bafa6e7af/merged major:0 minor:1270 fsType:overlay blockSize:0} overlay_0-1275:{mountpoint:/var/lib/containers/storage/overlay/8fbd1079e89622771aa48762d2df9a90ca5458fbc4bc6eb52afb114a7ff113b0/merged major:0 minor:1275 fsType:overlay blockSize:0} overlay_0-1277:{mountpoint:/var/lib/containers/storage/overlay/94325b32c03155e6ef018846a391328df9da0717ea8132d0e7961643f7223bab/merged major:0 minor:1277 fsType:overlay blockSize:0} overlay_0-1282:{mountpoint:/var/lib/containers/storage/overlay/bed4311941909b3ad638b8d30d6514f7a8887e0ed3154e618d2373c7d4418392/merged major:0 minor:1282 fsType:overlay blockSize:0} overlay_0-1287:{mountpoint:/var/lib/containers/storage/overlay/cc0e1cc741abccb954c4b02eeefe6a3cf6717b94ff0630987d58f8a416f8d576/merged major:0 minor:1287 fsType:overlay blockSize:0} overlay_0-129:{mountpoint:/var/lib/containers/storage/overlay/69d3b41aab4f992addc9aa3d6f15431647883b3c252fe16c21c0b872baf1204f/merged major:0 minor:129 fsType:overlay blockSize:0} overlay_0-1292:{mountpoint:/var/lib/containers/storage/overlay/b4c98f4da579ffdfab2d031165713851cfa47c12acae106f30d9e7fc5fb79889/merged major:0 minor:1292 fsType:overlay blockSize:0} overlay_0-1294:{mountpoint:/var/lib/containers/storage/overlay/59232c625cff8c8eee950eb1d6d0245a039b62a72af4b576662524a05a057916/merged major:0 minor:1294 fsType:overlay blockSize:0} overlay_0-1302:{mountpoint:/var/lib/containers/storage/overlay/39e363013beb9333530bb850a96613550eb914215bfe23cf7e52a5ed6966d8ae/merged major:0 minor:1302 fsType:overlay blockSize:0} overlay_0-131:{mountpoint:/var/lib/containers/storage/overlay/71e74b988ee1323c79a5fc15ad72123627de924888792d10c63849dd96a813c1/merged major:0 minor:131 fsType:overlay blockSize:0} overlay_0-1314:{mountpoint:/var/lib/containers/storage/overlay/dc5f6c3aef08d5742e005b092d49945cd7a8d65b7f87831472a879ff5568fafa/merged major:0 minor:1314 fsType:overlay blockSize:0} overlay_0-1316:{mountpoint:/var/lib/containers/storage/overlay/8cff965f824be8caf2ccd3616da396357bd87c4587c104ddedbe26fc53d1ef74/merged major:0 minor:1316 fsType:overlay blockSize:0} overlay_0-1321:{mountpoint:/var/lib/containers/storage/overlay/f444bf9a192a8e4bc5ff6841d4a01209627e00c1703272c01bc84095feb0b557/merged major:0 minor:1321 fsType:overlay blockSize:0} overlay_0-1327:{mountpoint:/var/lib/containers/storage/overlay/5f7ec38ac6575b2328a23fd360af7eef2412e28d7745c81b7226fcd40138b9a0/merged major:0 minor:1327 fsType:overlay blockSize:0} overlay_0-133:{mountpoint:/var/lib/containers/storage/overlay/daacf88f373af098aafa20a9a3771ec853b70408688f81c00b21bb58d99a4f73/merged major:0 minor:133 fsType:overlay blockSize:0} overlay_0-1347:{mountpoint:/var/lib/containers/storage/overlay/21f4aed5624198b526ae00a6f3af5c95f9491a9240dabc75f4b7576cffeff157/merged major:0 minor:1347 fsType:overlay blockSize:0} overlay_0-1349:{mountpoint:/var/lib/containers/storage/overlay/b463185d0e03fcb5424b27e98116f559594cbe22da1effede610139b3ce09aaf/merged major:0 minor:1349 fsType:overlay blockSize:0} overlay_0-1351:{mountpoint:/var/lib/containers/storage/overlay/af479cc928532446c124716305693b6906b20e63b7abfac8721cfd62a41e8eb8/merged major:0 minor:1351 fsType:overlay blockSize:0} overlay_0-1353:{mountpoint:/var/lib/containers/storage/overlay/56865bf890f73ec08b718b7352d4e736a9115244b2607f7c86795a5603291c19/merged major:0 minor:1353 fsType:overlay blockSize:0} overlay_0-1367:{mountpoint:/var/lib/containers/storage/overlay/e06334aa822566e8d00dbfac119cf42d0a85ba8056e3fbbe668802ad74d107ba/merged major:0 minor:1367 fsType:overlay blockSize:0} overlay_0-1376:{mountpoint:/var/lib/containers/storage/overlay/95c0e7fc72613eca789d0d6d31c23d375da87bbeac2de30e81a2a86e35127f08/merged major:0 minor:1376 fsType:overlay blockSize:0} overlay_0-1391:{mountpoint:/var/lib/containers/storage/overlay/1631fbfda6b5dcdd4871813b73986951985a76dd72130e89c2125a368332e10e/merged major:0 minor:1391 fsType:overlay blockSize:0} overlay_0-1399:{mountpoint:/var/lib/containers/storage/overlay/511597911d42bab5179e1cfa6664a7750261684f137665088667ea1c09d66261/merged major:0 minor:1399 fsType:overlay blockSize:0} overlay_0-1405:{mountpoint:/var/lib/containers/storage/overlay/c56f901ad542636dd1a1fde738f7988584f01f90c17f4fe47aec546d7256003b/merged major:0 minor:1405 fsType:overlay blockSize:0} overlay_0-1407:{mountpoint:/var/lib/containers/storage/overlay/ef2c7887b8699618702862bac4a4749afa0da5238a67e97673e6dd5738039fc9/merged major:0 minor:1407 fsType:overlay blockSize:0} overlay_0-1409:{mountpoint:/var/lib/containers/storage/overlay/334564cd0d1839f01bf344341b7497ec88da8e92049faec787d36ae4a95c9e71/merged major:0 minor:1409 fsType:overlay blockSize:0} overlay_0-141:{mountpoint:/var/lib/containers/storage/overlay/dca3e3b42802905a4dd825bcf333c36f4a2455d22e502d60213c75500a50bd69/merged major:0 minor:141 fsType:overlay blockSize:0} overlay_0-1434:{mountpoint:/var/lib/containers/storage/overlay/783fd3a5a4c5960ea4c0230937680c9a40f4374401d657adabdd89908d6b097a/merged major:0 minor:1434 fsType:overlay blockSize:0} overlay_0-1448:{mountpoint:/var/lib/containers/storage/overlay/1dff5aef1bb7bc4b1bab951e4df30296b0ace1c593fd4e2b6a8ea046ce5e21c1/merged major:0 minor:1448 fsType:overlay blockSize:0} overlay_0-1450:{mountpoint:/var/lib/containers/storage/overlay/07c8090737f9ecf2ed64d2b20c160bb1a6147046e3f336cc735c19d40a26ad37/merged major:0 minor:1450 fsType:overlay blockSize:0} overlay_0-1459:{mountpoint:/var/lib/containers/storage/overlay/8f50a0fe989bffdc4885610a2997ebdf6a7b86a70201b0ee5a698b3ce0ce9cad/merged major:0 minor:1459 fsType:overlay blockSize:0} overlay_0-147:{mountpoint:/var/lib/containers/storage/overlay/b0b75e0c84c872067c1cc7c0120ef370dfc2f8cb5b8b502f3235970ed83eb3c3/merged major:0 minor:147 fsType:overlay blockSize:0} overlay_0-1480:{mountpoint:/var/lib/containers/storage/overlay/04829e799367388de0ed4da094805ade49ea294cf7dc3980c9fe60805d49c7f9/merged major:0 minor:1480 fsType:overlay blockSize:0} overlay_0-1484:{mountpoint:/var/lib/containers/storage/overlay/3890023e872da73fc4a9e54cf40a23cd52bdd4e71ba5ffc5dc774ba8f15c8df3/merged major:0 minor:1484 fsType:overlay blockSize:0} overlay_0-1486:{mountpoint:/var/lib/containers/storage/overlay/b1a2659ee85c91d1442a68d7864177c2ab3f8ee7ed98196ccea016ceb8efbe73/merged major:0 minor:1486 fsType:overlay blockSize:0} overlay_0-1495:{mountpoint:/var/lib/containers/storage/overlay/32af3766ae2106202d930d82fc2004ecf41f4ca68c2316bd9c3c9ea1b24510ec/merged major:0 minor:1495 fsType:overlay blockSize:0} overlay_0-1497:{mountpoint:/var/lib/containers/storage/overlay/f17b1dfb9235b691152e93b615a6ceb739868ab0c801db4fb8eb515f54109fee/merged major:0 minor:1497 fsType:overlay blockSize:0} overlay_0-1505:{mountpoint:/var/lib/containers/storage/overlay/3e672cf9913761db51ec5a3257502bb3a5556b2d95bcaafb72ae4abdeba977d7/merged major:0 minor:1505 fsType:overlay blockSize:0} overlay_0-1507:{mountpoint:/var/lib/containers/storage/overlay/abc6cc35fa84163742cfb91d9a1eee26abe151e06ad590eb7933239b08776cda/merged major:0 minor:1507 fsType:overlay blockSize:0} overlay_0-1509:{mountpoint:/var/lib/containers/storage/overlay/abaf2cad0d13b6cafe4625ecf359dadb370ff884b55b111b34eeba64f2ce6a16/merged major:0 minor:1509 fsType:overlay blockSize:0} overlay_0-151:{mountpoint:/var/lib/containers/storage/overlay/ad73c9feb4b6df5992086ded295e9447fe56551a6bc2abfe15c347ec252b2df2/merged major:0 minor:151 fsType:overlay blockSize:0} overlay_0-1511:{mountpoint:/var/lib/containers/storage/overlay/794318cf92239c938d40949271145449430f611ea5928ab3ada947ec1dd9a54d/merged major:0 minor:1511 fsType:overlay blockSize:0} overlay_0-1532:{mountpoint:/var/lib/containers/storage/overlay/b7502c5bd0640c6f0a8afcd32de550dc46f81926026fc0bab92269fe96f72768/merged major:0 minor:1532 fsType:overlay blockSize:0} overlay_0-1534:{mountpoint:/var/lib/containers/storage/overlay/040dc743b34e28f9c89ba534defc42cb61c64758206789e36a953538e4331194/merged major:0 minor:1534 fsType:overlay blockSize:0} overlay_0-154:{mountpoint:/var/lib/containers/storage/overlay/ecb987277e86a099a527f281a809eca6a1e61d1361d335422ef901559beeac23/merged major:0 minor:154 fsType:overlay blockSize:0} overlay_0-1548:{mountpoint:/var/lib/containers/storage/overlay/9ed018b32b449301fdb32a5a7b7254ed17e06faedd72a3cf048b8a6f2a8aa5f7/merged major:0 minor:1548 fsType:overlay blockSize:0} overlay_0-156:{mountpoint:/var/lib/containers/storage/overlay/2adaf01a0b08d13626a4da6df7e18b5e4907881442b28b7d90d8833f5c3ac5e8/merged major:0 minor:156 fsType:overlay blockSize:0} overlay_0-1562:{mountpoint:/var/lib/containers/storage/overlay/e8c9fa8d72a09190b08d8bcb6edc9ead3f6a468174b3ec0e77fe9bccc71fe786/merged major:0 minor:1562 fsType:overlay blockSize:0} overlay_0-1564:{mountpoint:/var/lib/containers/storage/overlay/39f0df33b87554892fefd10a2de7abd130ff58a026d5d4afd3d0c8f051a9371b/merged major:0 minor:1564 fsType:overlay blockSize:0} overlay_0-1573:{mountpoint:/var/lib/containers/storage/overlay/464a07c041c94d36143931e4791a92a660dd186c954ab7b3c3372be5660102e4/merged major:0 minor:1573 fsType:overlay blockSize:0} overlay_0-1576:{mountpoint:/var/lib/containers/storage/overlay/169a14033f4635eeed279e3624d29f57f5996fb9d67eb58193a049047164171b/merged major:0 minor:1576 fsType:overlay blockSize:0} overlay_0-1577:{mountpoint:/var/lib/containers/storage/overlay/cc55986acc7aa5da1946c492bbc68d91bff3f4e9b1604e7e5527a1e740145ecf/merged major:0 minor:1577 fsType:overlay blockSize:0} overlay_0-1578:{mountpoint:/var/lib/containers/storage/overlay/f02a2064144c68644c81fb3f54e37c9eab9372346b00321ebc45a08b843a2fc4/merged major:0 minor:1578 fsType:overlay blockSize:0} overlay_0-158:{mountpoint:/var/lib/containers/storage/overlay/8042c20fe3cb599e0d51ab721f80c316642c0ca72e8c9a9fc7cec85eeec21221/merged major:0 minor:158 fsType:overlay blockSize:0} overlay_0-1582:{mountpoint:/var/lib/containers/storage/overlay/69a925de948886c70ea1dbda0f5949a55d94f0758a62ebb3b96e832b617b4450/merged major:0 minor:1582 fsType:overlay blockSize:0} overlay_0-1583:{mountpoint:/var/lib/containers/storage/overlay/b574518fd620ce20ecce860f3a52df3a53a620e187868b6567d0fede758b25dd/merged major:0 minor:1583 fsType:overlay blockSize:0} overlay_0-1595:{mountpoint:/var/lib/containers/storage/overlay/e9d7056b15d840443a32945e19a18a94056d7d3916d0c010db5acf4a82251eb8/merged major:0 minor:1595 fsType:overlay blockSize:0} overlay_0-1600:{mountpoint:/var/lib/containers/storage/overlay/7e60f2c7f4e4c0d5ef132959e91c9724fd445b67a7c1fdf10ce39e10b7dccf7d/merged major:0 minor:1600 fsType:overlay blockSize:0} overlay_0-1605:{mountpoint:/var/lib/containers/storage/overlay/5db9bf96559c6fc6631babf40507a8a58c7b19acab6fd38a0f009f6b2573eadc/merged major:0 minor:1605 fsType:overlay blockSize:0} overlay_0-1621:{mountpoint:/var/lib/containers/storage/overlay/62622c05d097d3a8d2b563b15a6a00755f2da217b6e6cdeb90d24d374956f55c/merged major:0 minor:1621 fsType:overlay blockSize:0} overlay_0-1641:{mountpoint:/var/lib/containers/storage/overlay/dcd99b0b915f176b03b1a4f0950c52579db257ffb6777594df8febaba0e1b082/merged major:0 minor:1641 fsType:overlay blockSize:0} overlay_0-1644:{mountpoint:/var/lib/containers/storage/overlay/91e928e566f8b845d92ca131a4ce8bfed980e8f4023ed5ec8c0312d717de8876/merged major:0 minor:1644 fsType:overlay blockSize:0} overlay_0-1646:{mountpoint:/var/lib/containers/storage/overlay/7a92fc3f8d95a340bbddbf10cc2177157e2e015e01caac0dbeb3253adc933aa1/merged major:0 minor:1646 fsType:overlay blockSize:0} overlay_0-1655:{mountpoint:/var/lib/containers/storage/overlay/d7794a3c2f1140f039d3aa750e6f93da9237f593f2da5c5cc3636bde1a959ced/merged major:0 minor:1655 fsType:overlay blockSize:0} overlay_0-167:{mountpoint:/var/lib/containers/storage/overlay/221c2f7e66b93d86e9da7232d411202d0694cf12b29cb712c4960a5e5e5c5024/merged major:0 minor:167 fsType:overlay blockSize:0} overlay_0-1671:{mountpoint:/var/lib/containers/storage/overlay/51eb5b8cfc467871ba56c1619118327f3c1f83dc8dcf3622de8660affc457f27/merged major:0 minor:1671 fsType:overlay blockSize:0} overlay_0-169:{mountpoint:/var/lib/containers/storage/overlay/c1f05399228e24704e4f3567767668c5f8597f886914f28b2747fe881ccf7121/merged major:0 minor:169 fsType:overlay blockSize:0} overlay_0-181:{mountpoint:/var/lib/containers/storage/overlay/c9fc3ab72976c7274a75cabc89076c5a3c76f05b2459d997d644b3fb0157cc5d/merged major:0 minor:181 fsType:overlay blockSize:0} overlay_0-183:{mountpoint:/var/lib/containers/storage/overlay/c8112fdb855c50fd639b9394dc2647a4c50f459f0efe7945c1ad234068025570/merged major:0 minor:183 fsType:overlay blockSize:0} overlay_0-185:{mountpoint:/var/lib/containers/storage/overlay/cd4f4481b0db7e5c34b29bfc60db4cceaf613a8470ee3c3c0eb9b6f37f70f133/merged major:0 minor:185 fsType:overlay blockSize:0} overlay_0-191:{mountpoint:/var/lib/containers/storage/overlay/672ba5764d28fa61ecdf4fa2dedbb811dd695a512500f55ce22fb456e4aaea22/merged major:0 minor:191 fsType:overlay blockSize:0} overlay_0-193:{mountpoint:/var/lib/containers/storage/overlay/44634517510c9242ca7a2ef11d48400d2344fe1b871f8b485681cfc8c11a0f75/merged major:0 minor:193 fsType:overlay blockSize:0} overlay_0-195:{mountpoint:/var/lib/containers/storage/overlay/d98934f7894d10a0de0cf8eaa27f3e1cf20717558b2da855607a3157b9a43913/merged major:0 minor:195 fsType:overlay blockSize:0} overlay_0-197:{mountpoint:/var/lib/containers/storage/overlay/0b6fc7c9d4ce32891e9527036aec70da87ba1aab65c0ee4ccd9d0a818a4fedb1/merged major:0 minor:197 fsType:overlay blockSize:0} overlay_0-199:{mountpoint:/var/lib/containers/storage/overlay/c700fbca17bf0aa18390fa0e7a3e7e50c95407e97b16040c11ff75c145e3024b/merged major:0 minor:199 fsType:overlay blockSize:0} overlay_0-201:{mountpoint:/var/lib/containers/storage/overlay/0e53020bc438850f0c0926f9b8c472725a6f9975b6eb735356a425f212d76b03/merged major:0 minor:201 fsType:overlay blockSize:0} overlay_0-203:{mountpoint:/var/lib/containers/storage/overlay/5e313736ef6becd5dc9dcb0dab1b97c8b5e156afb6bd893884068c1c498ce087/merged major:0 minor:203 fsType:overlay blockSize:0} overlay_0-205:{mountpoint:/var/lib/containers/storage/overlay/ee131bae2ad4ad92689b394d63362df84d3c96ef9aa039e40b427edb62c1d215/merged major:0 minor:205 fsType:overlay blockSize:0} overlay_0-209:{mountpoint:/var/lib/containers/storage/overlay/739cb662179f4420ea4e414cb5ccb3a85f1ebfdd8b84bd61ef7f9ddc35dc8308/merged major:0 minor:209 fsType:overlay blockSize:0} overlay_0-211:{mountpoint:/var/lib/containers/storage/overlay/fca1ec355e103423b556162062489b67d3d4e503bf4faaad9204b565468dedb6/merged major:0 minor:211 fsType:overlay blockSize:0} overlay_0-213:{mountpoint:/var/lib/containers/storage/overlay/746dae4fca92f1f46856320f3142b191f82009bfc78da76cf9934fe6c479eb21/merged major:0 minor:213 fsType:overlay blockSize:0} overlay_0-224:{mountpoint:/var/lib/containers/storage/overlay/489b045306615b560c94ca2d8cdd290d078e9f51d748acba0485b05bfb6def95/merged major:0 minor:224 fsType:overlay blockSize:0} overlay_0-228:{mountpoint:/var/lib/containers/storage/overlay/a0176a28a70d267f880769f37ef654ae085627da7e2a26813af8a468a87ad60f/merged major:0 minor:228 fsType:overlay blockSize:0} overlay_0-236:{mountpoint:/var/lib/containers/storage/overlay/f77b16658281bdf8a111ea497129031d7877cb11785b94d3092cb391f2202ca5/merged major:0 minor:236 fsType:overlay blockSize:0} overlay_0-244:{mountpoint:/var/lib/containers/storage/overlay/43ebf08663fa907da55be47d20f8b55b612ac6419332c76f1b701e88c5a677ae/merged major:0 minor:244 fsType:overlay blockSize:0} overlay_0-252:{mountpoint:/var/lib/containers/storage/overlay/99102a8dcf874c4ba63e522e1b531e1c54ff115e500f3b8f04ae20ce91113752/merged major:0 minor:252 fsType:overlay blockSize:0} overlay_0-268:{mountpoint:/var/lib/containers/storage/overlay/f11656f01663f1840822c3d5e2d7697ce24b7c255ab7bb1c54a68c7e6d64874b/merged major:0 minor:268 fsType:overlay blockSize:0} overlay_0-291:{mountpoint:/var/lib/containers/storage/overlay/c66d86781bfb63b46591c93693ca7996ba6b846502222a9ccb6778144909ebce/merged major:0 minor:291 fsType:overlay blockSize:0} overlay_0-294:{mountpoint:/var/lib/containers/storage/overlay/e8443993f3b7416ea6178a6b91bc11a40a839a0b6dd6475068141f140701c195/merged major:0 minor:294 fsType:overlay blockSize:0} overlay_0-316:{mountpoint:/var/lib/containers/storage/overlay/9e2277d70b2152dc3fcea05e7c7c9a5e5c37d900f125f6c37e400df22fa7a95b/merged major:0 minor:316 fsType:overlay blockSize:0} overlay_0-320:{mountpoint:/var/lib/containers/storage/overlay/6082bb1f87abff06cf3c105bba94e1ee8b30bed4bc88f2a11299cd54f691aa94/merged major:0 minor:320 fsType:overlay blockSize:0} overlay_0-322:{mountpoint:/var/lib/containers/storage/overlay/89b22bc61d954b1a593c909cbfe0436e763a79a3c3a22177c93dea7245201f66/merged major:0 minor:322 fsType:overlay blockSize:0} overlay_0-352:{mountpoint:/var/lib/containers/storage/overlay/07140a77fa648deb8b4d88c22e68fa10f6c7bac30b0c7786d3185b8459aeb839/merged major:0 minor:352 fsType:overlay blockSize:0} overlay_0-354:{mountpoint:/var/lib/containers/storage/overlay/49d8184721481ae0b07ee14398595a506ba2aae20f878baef545d290170ae9f1/merged major:0 minor:354 fsType:overlay blockSize:0} overlay_0-356:{mountpoint:/var/lib/containers/storage/overlay/0ccd97d1f30e293c96fca5436504b3210f4918b32e09318657425e960f17bc06/merged major:0 minor:356 fsType:overlay blockSize:0} overlay_0-358:{mountpoint:/var/lib/containers/storage/overlay/58b301a47cec9284ac1182393bdb11ccde7126768e1617ddb5f47b35e855e4dc/merged major:0 minor:358 fsType:overlay blockSize:0} overlay_0-360:{mountpoint:/var/lib/containers/storage/overlay/8c068bfa8aefed5fa7a6eaf9f0e7935b02af9100c42db54f0a8366f55dd75917/merged major:0 minor:360 fsType:overlay blockSize:0} overlay_0-362:{mountpoint:/var/lib/containers/storage/overlay/baeecbd481f89e1f70bbbe28bba0e050b3fe21bdfed072b3fc7cabdae408ba54/merged major:0 minor:362 fsType:overlay blockSize:0} overlay_0-364:{mountpoint:/var/lib/containers/storage/overlay/77eafadc8613b6f7fa1dd04e077d6685de20f2833bd5d2ecd5481f16ed6776cb/merged major:0 minor:364 fsType:overlay blockSize:0} overlay_0-366:{mountpoint:/var/lib/containers/storage/overlay/0b5b6e1e1402a79ebc6057fff4d379db33ef577dd917b6064f865035fa92d27e/merged major:0 minor:366 fsType:overlay blockSize:0} overlay_0-371:{mountpoint:/var/lib/containers/storage/overlay/8bd4f1bc77d295428965833bd172fb51fc966c0028405abc06c9f2478943788b/merged major:0 minor:371 fsType:overlay blockSize:0} overlay_0-373:{mountpoint:/var/lib/containers/storage/overlay/686f22baf944bf8c11d30dc16f26874154695a383a8cb5c54f403f2f36e2f393/merged major:0 minor:373 fsType:overlay blockSize:0} overlay_0-377:{mountpoint:/var/lib/containers/storage/overlay/15945a606966e8a5108e79db1845d2e9f5c51d24e304a2b03a319a76ced49309/merged major:0 minor:377 fsType:overlay blockSize:0} overlay_0-379:{mountpoint:/var/lib/containers/storage/overlay/d75c2056c63f1f756f7f527f91b3e94a861db48794706dd5523843f602ef7564/merged major:0 minor:379 fsType:overlay blockSize:0} overlay_0-381:{mountpoint:/var/lib/containers/storage/overlay/7246059058bd272c6b057dd7db6a7932b3512a067942e59be4995fcad774d800/merged major:0 minor:381 fsType:overlay blockSize:0} overlay_0-385:{mountpoint:/var/lib/containers/storage/overlay/d11123e76f1759aec07fd01ca6d1cd7946e7ce3359a0b20ed6518004832ddf75/merged major:0 minor:385 fsType:overlay blockSize:0} overlay_0-387:{mountpoint:/var/lib/containers/storage/overlay/855cc47dac6e9aeff40befed964a58516a945a51ec4b822c26f9f7e9ade252e5/merged major:0 minor:387 fsType:overlay blockSize:0} overlay_0-389:{mountpoint:/var/lib/containers/storage/overlay/195bde489fbe6cc048392f5e357b3240b8c98806e0da314a8f5c21e91a9bc2a8/merged major:0 minor:389 fsType:overlay blockSize:0} overlay_0-390:{mountpoint:/var/lib/containers/storage/overlay/918e0f320e7363ef658bcec35c9f6378d9f7164e94afa7773a4528ab3dac35e9/merged major:0 minor:390 fsType:overlay blockSize:0} overlay_0-392:{mountpoint:/var/lib/containers/storage/overlay/742c9fd3579942248a2b4f159fbe5a99b1e051eb9ff0418d64ab36c59b6a77b3/merged major:0 minor:392 fsType:overlay blockSize:0} overlay_0-394:{mountpoint:/var/lib/containers/storage/overlay/a8e6868bec75c94254f5cf8328fe0b44d096a58ac319b451fc29b333426b4146/merged major:0 minor:394 fsType:overlay blockSize:0} overlay_0-398:{mountpoint:/var/lib/containers/storage/overlay/9a30646052631ad7471a38651b056f6e4bfd683684942d114a770b463630e758/merged major:0 minor:398 fsType:overlay blockSize:0} overlay_0-400:{mountpoint:/var/lib/containers/storage/overlay/dc01f2c13de248e69d5c1a228ac808833202ed8e4d3f6182e161bcb146fb7f9c/merged major:0 minor:400 fsType:overlay blockSize:0} overlay_0-402:{mountpoint:/var/lib/containers/storage/overlay/163f2360566f2d7fa1a96361b76bd2218f5bd18089cca6c097c5b6a9e5d3f981/merged major:0 minor:402 fsType:overlay blockSize:0} overlay_0-404:{mountpoint:/var/lib/containers/storage/overlay/d08dd47ff25650be848a4aa97e3cde30bcf985640a213d05a8039f1b4ca3be9f/merged major:0 minor:404 fsType:overlay blockSize:0} overlay_0-406:{mountpoint:/var/lib/containers/storage/overlay/78b1844a69b54e740544612081f919ebd2d077f51b219b4ec0638fae6ed26ffe/merged major:0 minor:406 fsType:overlay blockSize:0} overlay_0-408:{mountpoint:/var/lib/containers/storage/overlay/9ac49532fb6a348bdb2072a72871b39f8bd086cf65f09fdeea58728b5ae29776/merged major:0 minor:408 fsType:overlay blockSize:0} overlay_0-410:{mountpoint:/var/lib/containers/storage/overlay/5d437a71bdb0c756d10ba5b0455fff4ada03aaddb2a941ef4e974dd41bbf32b2/merged major:0 minor:410 fsType:overlay blockSize:0} overlay_0-422:{mountpoint:/var/lib/containers/storage/overlay/9ad64f0c5c2b32d26ea88954bc46f1eb52680f649b06a5a343ccb5ccb3b64d19/merged major:0 minor:422 fsType:overlay blockSize:0} overlay_0-432:{mountpoint:/var/lib/containers/storage/overlay/06609e3fa92ceafad15371792b0131646fe9819d351872d0a5f45589c316a5f0/merged major:0 minor:432 fsType:overlay blockSize:0} overlay_0-434:{mountpoint:/var/lib/containers/storage/overlay/9aae3253efc3b2149f9833fbaa25fd3b2d94d02630b353422d3314fe3f19ea1b/merged major:0 minor:434 fsType:overlay blockSize:0} overlay_0-438:{mountpoint:/var/lib/containers/storage/overlay/8794a91423030a777c3699aa148ff7155805c5fa3795814617e9d8bff5eedca7/merged major:0 minor:438 fsType:overlay blockSize:0} overlay_0-44:{mountpoint:/var/lib/containers/storage/overlay/3c305c53cfd29952e6ca5d5fab57f55943a07ccb4b003a5beff65a97cd1655a4/merged major:0 minor:44 fsType:overlay blockSize:0} overlay_0-470:{mountpoint:/var/lib/containers/storage/overlay/39a618abcf5892e92a5c31e18efea6b29947cdf15eab61d5e1cc4c561da7187c/merged major:0 minor:470 fsType:overlay blockSize:0} overlay_0-474:{mountpoint:/var/lib/containers/storage/overlay/771d90bea5a5c152dcd6a3ba17dd084ac88cdf5f4426aa1741bd6ffd3a5d1bfb/merged major:0 minor:474 fsType:overlay blockSize:0} overlay_0-483:{mountpoint:/var/lib/containers/storage/overlay/9df50ee91ad13587e9da1d293b4056606b333ac5520288725c0864b2edb04538/merged major:0 minor:483 fsType:overlay blockSize:0} overlay_0-487:{mountpoint:/var/lib/containers/storage/overlay/da5c0507195eb3e459338c8f665b1b3f22beeab0a3a6187f6685895a0341a59c/merged major:0 minor:487 fsType:overlay Dec 05 10:47:20.073231 master-0 kubenswrapper[24928]: blockSize:0} overlay_0-493:{mountpoint:/var/lib/containers/storage/overlay/1bd39c1fbad74915ef709b10b8a45f024c0e3dbe39f6f4a52045aebfa2ea48b0/merged major:0 minor:493 fsType:overlay blockSize:0} overlay_0-496:{mountpoint:/var/lib/containers/storage/overlay/2f485396a0f509a687825c9f76c400f8730dee9ca0cc46243b15fdc35078e395/merged major:0 minor:496 fsType:overlay blockSize:0} overlay_0-499:{mountpoint:/var/lib/containers/storage/overlay/8ed98e690ce837871c4609503406edb0bb32e67bb389b1d82aa9d9c9887825a9/merged major:0 minor:499 fsType:overlay blockSize:0} overlay_0-501:{mountpoint:/var/lib/containers/storage/overlay/cec265e4df977ed96ad30b26dcc6beba93c97c2a2a03c249843c6f7c89579954/merged major:0 minor:501 fsType:overlay blockSize:0} overlay_0-511:{mountpoint:/var/lib/containers/storage/overlay/004ccc8b6dd081f01b82fa983ad06df05601006dfa32df59a2ee4f9f20664686/merged major:0 minor:511 fsType:overlay blockSize:0} overlay_0-513:{mountpoint:/var/lib/containers/storage/overlay/d19ba502a238b90a53d536b5043875c947041a33a7367d998f0e4bcd26a8e317/merged major:0 minor:513 fsType:overlay blockSize:0} overlay_0-527:{mountpoint:/var/lib/containers/storage/overlay/48403893120fc4f8f3f49bb98457d746b199a8314c3cea9626af3a5138bef5b4/merged major:0 minor:527 fsType:overlay blockSize:0} overlay_0-529:{mountpoint:/var/lib/containers/storage/overlay/cf87fc63dc8d895db2dd045aa4ea78cb37189274278f3845662eeac5be7acf72/merged major:0 minor:529 fsType:overlay blockSize:0} overlay_0-545:{mountpoint:/var/lib/containers/storage/overlay/66863eb2574b19b1eb25144a816f99cbd90bcea8cfb189b41114fd3694d581c8/merged major:0 minor:545 fsType:overlay blockSize:0} overlay_0-547:{mountpoint:/var/lib/containers/storage/overlay/4284847fbfb3d02f4dee6addb29d96bee7b0ba76dc0aa00fba176a079580f7d6/merged major:0 minor:547 fsType:overlay blockSize:0} overlay_0-559:{mountpoint:/var/lib/containers/storage/overlay/a7c2943f1921a3bddaee3d6981b17b1c1bdc51a1cbe897b9a9f15e3165bc7987/merged major:0 minor:559 fsType:overlay blockSize:0} overlay_0-560:{mountpoint:/var/lib/containers/storage/overlay/2576e8e0fb317d10369127bc57d0241a28abe9123c2e4b63b7b43d28ca94ff39/merged major:0 minor:560 fsType:overlay blockSize:0} overlay_0-566:{mountpoint:/var/lib/containers/storage/overlay/b3e15dded8b943e62351190f690eee2269e455dc15d6894e7065fad961ab85f7/merged major:0 minor:566 fsType:overlay blockSize:0} overlay_0-568:{mountpoint:/var/lib/containers/storage/overlay/91c366671b05d7ba1322ef3dcc01d0b9150402c8bc19aa3e86e8d6a2050ed025/merged major:0 minor:568 fsType:overlay blockSize:0} overlay_0-57:{mountpoint:/var/lib/containers/storage/overlay/98e30cfe4c56429a5772698b2e25c3ee80fe24470a485325df407654fe36967a/merged major:0 minor:57 fsType:overlay blockSize:0} overlay_0-570:{mountpoint:/var/lib/containers/storage/overlay/2805df90dc26b9832087815f4cc4448bea13fd5f3d1a0450d5669d303b0dccb6/merged major:0 minor:570 fsType:overlay blockSize:0} overlay_0-577:{mountpoint:/var/lib/containers/storage/overlay/437b1390d64e9104eac43feba3eab9798cac2ece77c4bde3e69e1977ef58705d/merged major:0 minor:577 fsType:overlay blockSize:0} overlay_0-58:{mountpoint:/var/lib/containers/storage/overlay/f55372796519f8e3240a50e2402021e2680925bdc11299969bc13c87b0404c46/merged major:0 minor:58 fsType:overlay blockSize:0} overlay_0-586:{mountpoint:/var/lib/containers/storage/overlay/27ea85602d1116ba2b6ad24a04dcd73b3c914f4e3921b42ac2bea5132258c192/merged major:0 minor:586 fsType:overlay blockSize:0} overlay_0-588:{mountpoint:/var/lib/containers/storage/overlay/ff27835a300e257efe664b72cd35293582303e1aaa0e602f8d6b15ef48d42b11/merged major:0 minor:588 fsType:overlay blockSize:0} overlay_0-595:{mountpoint:/var/lib/containers/storage/overlay/c03cd9240337973dd415677a646bc7424444dcc24944c9a470cdebeb31eff092/merged major:0 minor:595 fsType:overlay blockSize:0} overlay_0-60:{mountpoint:/var/lib/containers/storage/overlay/86931a7c9ecfa606bee4e6328d7ff24df409d3d237f58a76a79b260050de5de9/merged major:0 minor:60 fsType:overlay blockSize:0} overlay_0-601:{mountpoint:/var/lib/containers/storage/overlay/a0a18645c23e03c88993a93dbf3f41ed9f6e8d47fcbb5252ced18d16717334e8/merged major:0 minor:601 fsType:overlay blockSize:0} overlay_0-604:{mountpoint:/var/lib/containers/storage/overlay/8b6121f23249b3e42408b58290f957884ee627bee49203e8094b395851bc91b7/merged major:0 minor:604 fsType:overlay blockSize:0} overlay_0-606:{mountpoint:/var/lib/containers/storage/overlay/986274bb6ad0fbe6c43a718eefa241274b081209a42cd14f88a6b1c25caf4525/merged major:0 minor:606 fsType:overlay blockSize:0} overlay_0-618:{mountpoint:/var/lib/containers/storage/overlay/d4fce8290e2ef2f973bb41d571191c4e73bb4369860b06354ba119b264b568be/merged major:0 minor:618 fsType:overlay blockSize:0} overlay_0-62:{mountpoint:/var/lib/containers/storage/overlay/e53497d94773e663d8dec87fb79f8ea0aa7dd58034c7327f49d504187e4bfacc/merged major:0 minor:62 fsType:overlay blockSize:0} overlay_0-631:{mountpoint:/var/lib/containers/storage/overlay/7facc2daeb11b414e8e082bf22272412edef714ed6b577414b3ccb10878b6491/merged major:0 minor:631 fsType:overlay blockSize:0} overlay_0-633:{mountpoint:/var/lib/containers/storage/overlay/69848bf83ef34119535ebafb9b3faa193975f83e00889e1fa8e74ee92fe68a7f/merged major:0 minor:633 fsType:overlay blockSize:0} overlay_0-637:{mountpoint:/var/lib/containers/storage/overlay/39a2e9c20244bb2fd89f115bf9bdc40bfb85996ae74482bcc531aeb0ba6c2b39/merged major:0 minor:637 fsType:overlay blockSize:0} overlay_0-641:{mountpoint:/var/lib/containers/storage/overlay/9012ea0ae11ce855932c89ebba887581a2631bef2385bc8bcbba67ed48ddbf35/merged major:0 minor:641 fsType:overlay blockSize:0} overlay_0-650:{mountpoint:/var/lib/containers/storage/overlay/bd958ba46041d6b358da7206655c26b94dbb97e65b1a0a6279e94d9fd9495302/merged major:0 minor:650 fsType:overlay blockSize:0} overlay_0-652:{mountpoint:/var/lib/containers/storage/overlay/461bcd485d1b07a11d2ff40d92b0d002c497f48920016b6bda8c7a9cc1549267/merged major:0 minor:652 fsType:overlay blockSize:0} overlay_0-653:{mountpoint:/var/lib/containers/storage/overlay/71d2f2bd8ce40ef9bb0d409b192498569663f51ce3514702450067014a7579a1/merged major:0 minor:653 fsType:overlay blockSize:0} overlay_0-655:{mountpoint:/var/lib/containers/storage/overlay/c63ed3e6d7b4577f2af92952c6c942b2f6391126238531012e7eb083dc48a039/merged major:0 minor:655 fsType:overlay blockSize:0} overlay_0-66:{mountpoint:/var/lib/containers/storage/overlay/ee779b66a1c34e9457658cfd27e9e31743bab2d8d3e16c02edd27b86be3bbeb0/merged major:0 minor:66 fsType:overlay blockSize:0} overlay_0-674:{mountpoint:/var/lib/containers/storage/overlay/56e097200827a3245453e7a579b39033ab16ff8c71578fe86b2d5fbb990bf619/merged major:0 minor:674 fsType:overlay blockSize:0} overlay_0-687:{mountpoint:/var/lib/containers/storage/overlay/f7918c827ca26ce22ec9274920607318b063da3d293217f07adbb1712eb6eec6/merged major:0 minor:687 fsType:overlay blockSize:0} overlay_0-691:{mountpoint:/var/lib/containers/storage/overlay/20c77558415fb8ccacf405c1e51a8cfbffde634541886cc282d389f44aa56ddf/merged major:0 minor:691 fsType:overlay blockSize:0} overlay_0-693:{mountpoint:/var/lib/containers/storage/overlay/380170e91ba31eb654ceee5cd6f1b56532bf929c89f1951e3a1b1f55337d0425/merged major:0 minor:693 fsType:overlay blockSize:0} overlay_0-695:{mountpoint:/var/lib/containers/storage/overlay/3f30a9149fe134778a30251f7fcc40be9a3f5df7f335fe9447c3d3f1b06414fc/merged major:0 minor:695 fsType:overlay blockSize:0} overlay_0-697:{mountpoint:/var/lib/containers/storage/overlay/af9b29d96fbf108794f0570cc70f09f0bc2d9c7a242684c44b8cb07324aee954/merged major:0 minor:697 fsType:overlay blockSize:0} overlay_0-705:{mountpoint:/var/lib/containers/storage/overlay/f827a7faeea8fcdcb02190f61f183b1d0b076e1680091943b3c9cc55bf9879cb/merged major:0 minor:705 fsType:overlay blockSize:0} overlay_0-711:{mountpoint:/var/lib/containers/storage/overlay/ae20fc8d5703b5a108bc161d69134c3df954ad53900e56575b2292d1df5527db/merged major:0 minor:711 fsType:overlay blockSize:0} overlay_0-72:{mountpoint:/var/lib/containers/storage/overlay/a5815feb412038b92dc56e87cfdf2468519cbea4e1ba60c681f861c783ee51d3/merged major:0 minor:72 fsType:overlay blockSize:0} overlay_0-723:{mountpoint:/var/lib/containers/storage/overlay/0b955450d983cb251a0dff8d5a46351b15675e4e7f0d727b82fc69e56110b9f4/merged major:0 minor:723 fsType:overlay blockSize:0} overlay_0-725:{mountpoint:/var/lib/containers/storage/overlay/665c54623d3e5a3cac25073432b6962546b847051fde7d6419a2cb6503902683/merged major:0 minor:725 fsType:overlay blockSize:0} overlay_0-729:{mountpoint:/var/lib/containers/storage/overlay/5b77a165697ff5c5ba2ceaae836e83eca7d3fc69056eacd3f9e1248c63ab7480/merged major:0 minor:729 fsType:overlay blockSize:0} overlay_0-731:{mountpoint:/var/lib/containers/storage/overlay/0c1d2e7826f29815f4b3b118d2eeacad9098bc9cede7468856a8571fb537355d/merged major:0 minor:731 fsType:overlay blockSize:0} overlay_0-744:{mountpoint:/var/lib/containers/storage/overlay/b5d2559e09e6250708a3d2364a4f19f4bffc97e27a6e3cb1d895c9881fb297a8/merged major:0 minor:744 fsType:overlay blockSize:0} overlay_0-747:{mountpoint:/var/lib/containers/storage/overlay/25cd27e44881baf3f27857e2460936474f23474a8eb23e137c6a58135db13959/merged major:0 minor:747 fsType:overlay blockSize:0} overlay_0-749:{mountpoint:/var/lib/containers/storage/overlay/1313596ade2637e1fa2b20bcc54fc0826c0da8d30fc4eae08821c1c87512a3af/merged major:0 minor:749 fsType:overlay blockSize:0} overlay_0-76:{mountpoint:/var/lib/containers/storage/overlay/a3336bc13758cef0a6e91cbd627f2d88b4568114095e1579405b43a0487c668f/merged major:0 minor:76 fsType:overlay blockSize:0} overlay_0-77:{mountpoint:/var/lib/containers/storage/overlay/ab22c55f4db9dd1d2a9a0048cb9a7b0af2982e12c503c48ce80d67288299873d/merged major:0 minor:77 fsType:overlay blockSize:0} overlay_0-770:{mountpoint:/var/lib/containers/storage/overlay/caafb84fb88fb9c95cfa88cae3fb845da6ac7c0cf8129b81b7ebd7c68f3c57e7/merged major:0 minor:770 fsType:overlay blockSize:0} overlay_0-786:{mountpoint:/var/lib/containers/storage/overlay/712a5812a269236dbd62f0dd6dee1b382a2c39002b2a3bf9ca8d16a6eba2ebdf/merged major:0 minor:786 fsType:overlay blockSize:0} overlay_0-788:{mountpoint:/var/lib/containers/storage/overlay/1e0be391cd575e3efeff8b069ab8a56ffa641e99093a998566c3f6ec1e415d9f/merged major:0 minor:788 fsType:overlay blockSize:0} overlay_0-789:{mountpoint:/var/lib/containers/storage/overlay/d8be5a63e6e9b547b3b5816e81e7cd70b6933643ce855d6a8d2f32b490e8f386/merged major:0 minor:789 fsType:overlay blockSize:0} overlay_0-790:{mountpoint:/var/lib/containers/storage/overlay/17e5917bf80d22ce633631af656fd1aa13f5c154debf8fc6f8da4fa776c8181d/merged major:0 minor:790 fsType:overlay blockSize:0} overlay_0-792:{mountpoint:/var/lib/containers/storage/overlay/983999ac5c481f5c3c8e7b43fb806439905caca395a550219eb634255a95b1a2/merged major:0 minor:792 fsType:overlay blockSize:0} overlay_0-803:{mountpoint:/var/lib/containers/storage/overlay/0a36b59e2f9efd7f9e3a924d683df61acdffdb3eb6b3e25df99fe25560eb4eb9/merged major:0 minor:803 fsType:overlay blockSize:0} overlay_0-822:{mountpoint:/var/lib/containers/storage/overlay/c9ec1f2fadfbd7975128517d07ac631e0d1f6b103f7946c0c17e29b055e20dcf/merged major:0 minor:822 fsType:overlay blockSize:0} overlay_0-824:{mountpoint:/var/lib/containers/storage/overlay/f797a28cb32143d3d2a29984e0734082f7b70644e0377ef53631a05c12d445e7/merged major:0 minor:824 fsType:overlay blockSize:0} overlay_0-826:{mountpoint:/var/lib/containers/storage/overlay/36a327594fd2a2bab630e77110795ddb43dc8616c4309f4e99319f903ef08518/merged major:0 minor:826 fsType:overlay blockSize:0} overlay_0-83:{mountpoint:/var/lib/containers/storage/overlay/15d5eebe9f641d2a2bcf1349d7b28a7bad665a7b12c03e07ff17463b310b780b/merged major:0 minor:83 fsType:overlay blockSize:0} overlay_0-846:{mountpoint:/var/lib/containers/storage/overlay/6c47c81e000b7f29f62135ebeb3ddd8f7054eb39f0d3a0514c1b5ff60f612097/merged major:0 minor:846 fsType:overlay blockSize:0} overlay_0-848:{mountpoint:/var/lib/containers/storage/overlay/ce21a2f8c8eee8d17ab2fd86717317b2dc3032c8cab9342b9c372965450fc3b7/merged major:0 minor:848 fsType:overlay blockSize:0} overlay_0-850:{mountpoint:/var/lib/containers/storage/overlay/0836ad524e4506211e59b3215050b43bb932a149f666e28342368a866b2822cf/merged major:0 minor:850 fsType:overlay blockSize:0} overlay_0-861:{mountpoint:/var/lib/containers/storage/overlay/4b9e63b0556eb34d55aac3e12d8a7eecd9154af7af7235f3427fd1529390675a/merged major:0 minor:861 fsType:overlay blockSize:0} overlay_0-863:{mountpoint:/var/lib/containers/storage/overlay/b88fc8e4cb3532c0743cbe12497cf904db483e808fb48a75235524acda1f3d26/merged major:0 minor:863 fsType:overlay blockSize:0} overlay_0-867:{mountpoint:/var/lib/containers/storage/overlay/34c226abc160a78a0d33aaa844008b0156d0c4090171643ded2812f3c6c6c860/merged major:0 minor:867 fsType:overlay blockSize:0} overlay_0-877:{mountpoint:/var/lib/containers/storage/overlay/a3519242422eddfa33f9578fd1f956d5445663c18acbf3cf2c37264dd2ddc977/merged major:0 minor:877 fsType:overlay blockSize:0} overlay_0-881:{mountpoint:/var/lib/containers/storage/overlay/57a5b7a2441f345c045649805537d06bb4dc6e1c0e2dfa46895da584f61d99d3/merged major:0 minor:881 fsType:overlay blockSize:0} overlay_0-90:{mountpoint:/var/lib/containers/storage/overlay/5c26d1bfa249e7c6c3f330231cbbfb3ad947177997283c8508831b6e9d3152a9/merged major:0 minor:90 fsType:overlay blockSize:0} overlay_0-902:{mountpoint:/var/lib/containers/storage/overlay/c7e10a0daa30633fe8b1bd228fca8fc7161f62786af9f055083b9ea66c366eff/merged major:0 minor:902 fsType:overlay blockSize:0} overlay_0-907:{mountpoint:/var/lib/containers/storage/overlay/23bf5ee695f465f0dae4d30b46872daf6349bb2ee7d158541ef464aae2c5bf4a/merged major:0 minor:907 fsType:overlay blockSize:0} overlay_0-909:{mountpoint:/var/lib/containers/storage/overlay/aecbaa86fabecc6ea209ae1c12da01d259b3b15787152861a006ff38965e0684/merged major:0 minor:909 fsType:overlay blockSize:0} overlay_0-911:{mountpoint:/var/lib/containers/storage/overlay/77b739baa81a8224f74772e11374408b0412e201099336b05dd53e79dbef6f3d/merged major:0 minor:911 fsType:overlay blockSize:0} overlay_0-913:{mountpoint:/var/lib/containers/storage/overlay/e51c4cc5dd297ff11113137f95c72e066bc9afb261ee9f4351b7aebd45f8dc7b/merged major:0 minor:913 fsType:overlay blockSize:0} overlay_0-92:{mountpoint:/var/lib/containers/storage/overlay/05bb803fed1ffb632eaf81441872a2ecbd71a83534cd0461518018d34e8abaf8/merged major:0 minor:92 fsType:overlay blockSize:0} overlay_0-93:{mountpoint:/var/lib/containers/storage/overlay/c58617f5fc0a025fda9bff765bcb66873a417fcd95e46ea6adbb4ac7a7587de7/merged major:0 minor:93 fsType:overlay blockSize:0} overlay_0-939:{mountpoint:/var/lib/containers/storage/overlay/80f310e3d59e410009609fced4ce092e9a7108a4a7738ed2be14e142075146a0/merged major:0 minor:939 fsType:overlay blockSize:0} overlay_0-941:{mountpoint:/var/lib/containers/storage/overlay/54a44e5421c20307f85052e5bf2cbdda4f11c4f877c23eadef5c1a16e0bc6a68/merged major:0 minor:941 fsType:overlay blockSize:0} overlay_0-943:{mountpoint:/var/lib/containers/storage/overlay/046c04b66dac966c6cb8d183c3d57205bfc256c0c1653dcf2246582793f199a2/merged major:0 minor:943 fsType:overlay blockSize:0} overlay_0-945:{mountpoint:/var/lib/containers/storage/overlay/4f9565a9e9bd1714980d103b75b809eea8b6068c3012d9dbdcff41d613a454bc/merged major:0 minor:945 fsType:overlay blockSize:0} overlay_0-959:{mountpoint:/var/lib/containers/storage/overlay/d080de9640f46917d44362f933398f6a047ce134a77b5044c071e45f44108f08/merged major:0 minor:959 fsType:overlay blockSize:0} overlay_0-961:{mountpoint:/var/lib/containers/storage/overlay/3e6b3b4474c9a4b4534235cb7c5cfbaa6b60eb115e3fbf6cc7f1f97ab2da9306/merged major:0 minor:961 fsType:overlay blockSize:0} overlay_0-97:{mountpoint:/var/lib/containers/storage/overlay/fedb74a2fa4d647b240d85c5830e55dc6cc5cd0233341e58ae8e85f31c856b65/merged major:0 minor:97 fsType:overlay blockSize:0} overlay_0-972:{mountpoint:/var/lib/containers/storage/overlay/d6d7b204a21201a2586a208da08566001fbfedbc163a5223ab04ef068e4eda7c/merged major:0 minor:972 fsType:overlay blockSize:0} overlay_0-981:{mountpoint:/var/lib/containers/storage/overlay/fdcebce4dfa78d7009eba909fe38f807175681d60ed926d4dc605fb28b99c8c1/merged major:0 minor:981 fsType:overlay blockSize:0} overlay_0-990:{mountpoint:/var/lib/containers/storage/overlay/4c49d732335146c3962bf93697c76c94051406823bf9cfc1b4ffec2852375e58/merged major:0 minor:990 fsType:overlay blockSize:0} overlay_0-992:{mountpoint:/var/lib/containers/storage/overlay/ecce0c92538abf19ceb8259663761f7a9448eaab7bbfb01609406dfb81131be7/merged major:0 minor:992 fsType:overlay blockSize:0} overlay_0-995:{mountpoint:/var/lib/containers/storage/overlay/bb1803c25779a331b81c2e184cbcd1a66d08b4754fe1863b55b28158394b57cf/merged major:0 minor:995 fsType:overlay blockSize:0} overlay_0-997:{mountpoint:/var/lib/containers/storage/overlay/a38c6062a74cb523949e8ce8a83f39a322406e2fef9a844e9fa836a1e5acb46d/merged major:0 minor:997 fsType:overlay blockSize:0} overlay_0-999:{mountpoint:/var/lib/containers/storage/overlay/523627bd2c8d691dd10690e8ff2ccee85e784cb89af91b7bd1ec6575179f3a14/merged major:0 minor:999 fsType:overlay blockSize:0}] Dec 05 10:47:20.118228 master-0 kubenswrapper[24928]: I1205 10:47:20.116899 24928 manager.go:217] Machine: {Timestamp:2025-12-05 10:47:20.116017251 +0000 UTC m=+0.119211122 CPUVendorID:AuthenticAMD NumCores:16 NumPhysicalCores:1 NumSockets:16 CpuFrequency:2799998 MemoryCapacity:50514153472 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:9319397e1208473fb2e5a8b942c60cc5 SystemUUID:9319397e-1208-473f-b2e5-a8b942c60cc5 BootID:759865f8-012e-4147-b61d-8ab8adc2639f Filesystems:[{Device:/var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/bound-sa-token DeviceMajor:0 DeviceMinor:312 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1302 DeviceMajor:0 DeviceMinor:1302 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/21a9bd5a-3ade-48ef-8004-e0492daa85ba/volumes/kubernetes.io~secret/node-exporter-tls DeviceMajor:0 DeviceMinor:1470 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1006 DeviceMajor:0 DeviceMinor:1006 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-58 DeviceMajor:0 DeviceMinor:58 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/88cccb5b-1ad1-4fab-b34c-90252794ee20/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:952 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1082 DeviceMajor:0 DeviceMinor:1082 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1605 DeviceMajor:0 DeviceMinor:1605 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:259 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-320 DeviceMajor:0 DeviceMinor:320 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/ccbec542fddb9bd0831349589c4ac65650d79b1cf75a0af75cf8ca2d1a0d89f6/userdata/shm DeviceMajor:0 DeviceMinor:337 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-577 DeviceMajor:0 DeviceMinor:577 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-169 DeviceMajor:0 DeviceMinor:169 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/0a4db531-8af7-4085-a6b2-0de51b527ce6/volumes/kubernetes.io~projected/kube-api-access-g4jgc DeviceMajor:0 DeviceMinor:1171 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/4cedf7a9bbcb2d72fff4df0c2eb3dda7d05531d0220b30c3f687e88816b7ce46/userdata/shm DeviceMajor:0 DeviceMinor:1478 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/4c76b1c660e01ccc1ed121db81862ca5648ff5a628cbb72bf8b46891908897a5/userdata/shm DeviceMajor:0 DeviceMinor:1172 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1548 DeviceMajor:0 DeviceMinor:1548 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:265 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~projected/kube-api-access-7wfsv DeviceMajor:0 DeviceMinor:282 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-224 DeviceMajor:0 DeviceMinor:224 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1080 DeviceMajor:0 DeviceMinor:1080 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~projected/kube-api-access-jv4zs DeviceMajor:0 DeviceMinor:286 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/97192d4d-15d3-4740-82f3-d0d45f9fe7b9/volumes/kubernetes.io~secret/signing-key DeviceMajor:0 DeviceMinor:498 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/01488ac4-313d-48d5-9e86-7b2011b9e91e/volumes/kubernetes.io~secret/proxy-tls DeviceMajor:0 DeviceMinor:1044 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-788 DeviceMajor:0 DeviceMinor:788 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-687 DeviceMajor:0 DeviceMinor:687 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-911 DeviceMajor:0 DeviceMinor:911 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-373 DeviceMajor:0 DeviceMinor:373 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/6728a458-098b-44f9-8c8f-b0d76c5825fa/volumes/kubernetes.io~projected/kube-api-access-784mb DeviceMajor:0 DeviceMinor:1311 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-501 DeviceMajor:0 DeviceMinor:501 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-201 DeviceMajor:0 DeviceMinor:201 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-211 DeviceMajor:0 DeviceMinor:211 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~projected/kube-api-access-5rlts DeviceMajor:0 DeviceMinor:1559 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~secret/metrics-tls DeviceMajor:0 DeviceMinor:43 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~projected/kube-api-access-6vp9b DeviceMajor:0 DeviceMinor:289 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/8c649a16-c187-412e-b5da-62a00bee38ab/volumes/kubernetes.io~projected/kube-api-access-d8pv2 DeviceMajor:0 DeviceMinor:309 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/baee05cb-62ad-4eda-8a13-e317eae840e5/volumes/kubernetes.io~projected/kube-api-access-564nt DeviceMajor:0 DeviceMinor:726 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1562 DeviceMajor:0 DeviceMinor:1562 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/ee0b4a7f-1b96-4304-bef0-fb575e0e431c/volumes/kubernetes.io~projected/kube-api-access-9ktnb DeviceMajor:0 DeviceMinor:1053 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-992 DeviceMajor:0 DeviceMinor:992 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1327 DeviceMajor:0 DeviceMinor:1327 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~secret/metrics-certs DeviceMajor:0 DeviceMinor:1336 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/e2506cc6607d77e49135fcb375a37bafe2d511f6ed75718677873e40b092ab87/userdata/shm DeviceMajor:0 DeviceMinor:333 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/5ab705ed0c0dfea5ee8845410f03f5084071123b3566b80a7ef7e6fcce0fa974/userdata/shm DeviceMajor:0 DeviceMinor:383 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1095 DeviceMajor:0 DeviceMinor:1095 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/d996cd9dee5c72648a9d435634d8c36eb7feb019522350be8d7d8f4a1fa79c46/userdata/shm DeviceMajor:0 DeviceMinor:1482 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/0d2211dcab91af9edc388c398c5fc7a90c2e8c65270b4ebc9dead574b96eaefe/userdata/shm DeviceMajor:0 DeviceMinor:332 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/ba6a43bb61d242c8e032f7cab8a9cf0e806cdf03c0a9ce44ec6c444a2216f39a/userdata/shm DeviceMajor:0 DeviceMinor:624 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1158 DeviceMajor:0 DeviceMinor:1158 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1181 DeviceMajor:0 DeviceMinor:1181 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1497 DeviceMajor:0 DeviceMinor:1497 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/2cf0d485e10f74fdd9bf0282b8f6fbc341fecfb79764618939e754565cc0b6ff/userdata/shm DeviceMajor:0 DeviceMinor:1076 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1405 DeviceMajor:0 DeviceMinor:1405 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/b7452830de99c0487a8d247b23da92391c3a1113eb379e2b40ac1673b8cb1fe1/userdata/shm DeviceMajor:0 DeviceMinor:336 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/fec7b5fa2a30a341bff8e373ec4fbcd19900103be9e9af536060bcabaccdcce1/userdata/shm DeviceMajor:0 DeviceMinor:727 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/1fd0a349-0f66-4c85-95e4-8bbc96648c05/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:883 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/88cccb5b-1ad1-4fab-b34c-90252794ee20/volumes/kubernetes.io~projected/kube-api-access-f4lbg DeviceMajor:0 DeviceMinor:953 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-244 DeviceMajor:0 DeviceMinor:244 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-822 DeviceMajor:0 DeviceMinor:822 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-108 DeviceMajor:0 DeviceMinor:108 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/b7e7f216-f9af-41df-a0b0-df2a76b9b72a/volumes/kubernetes.io~secret/proxy-tls DeviceMajor:0 DeviceMinor:1169 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1511 DeviceMajor:0 DeviceMinor:1511 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/b94fba3ebe8eacc577514add51b89ec37946322bc96bbe0ee6d3e8d5b7830ae3/userdata/shm DeviceMajor:0 DeviceMinor:1369 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/8d62f303e35a44bd684cd9b7f51835e4e80959496c433fcc66b39cdd84c84cc8/userdata/shm DeviceMajor:0 DeviceMinor:149 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-770 DeviceMajor:0 DeviceMinor:770 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-106 DeviceMajor:0 DeviceMinor:106 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1495 DeviceMajor:0 DeviceMinor:1495 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/bound-sa-token DeviceMajor:0 DeviceMinor:308 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-66 DeviceMajor:0 DeviceMinor:66 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/5046caa4441a1760a91554be78c89b5b216c5aae3f10502f03a571a1cc07afd4/userdata/shm DeviceMajor:0 DeviceMinor:1069 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:25257074688 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/4c2975ec-e33d-4960-a708-277d41c79b15/volumes/kubernetes.io~projected/kube-api-access-d6z8n DeviceMajor:0 DeviceMinor:1046 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1277 DeviceMajor:0 DeviceMinor:1277 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/7e22b55ec9c49898cca9f69ea15552e8a230cbb6781fbf80f0e4919b22b7b7d8/userdata/shm DeviceMajor:0 DeviceMinor:338 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/a4fd453c-e667-4bdc-aa9e-3d95ff707200/volumes/kubernetes.io~secret/catalogserver-certs DeviceMajor:0 DeviceMinor:815 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-493 DeviceMajor:0 DeviceMinor:493 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-595 DeviceMajor:0 DeviceMinor:595 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1148 DeviceMajor:0 DeviceMinor:1148 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1450 DeviceMajor:0 DeviceMinor:1450 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7/userdata/shm DeviceMajor:0 DeviceMinor:41 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-252 DeviceMajor:0 DeviceMinor:252 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:277 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-352 DeviceMajor:0 DeviceMinor:352 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/1a0235af-2cf2-4ad4-b419-764fb56a0107/volumes/kubernetes.io~projected/kube-api-access-cwqkb DeviceMajor:0 DeviceMinor:122 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-362 DeviceMajor:0 DeviceMinor:362 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/d9ff1ae2-8547-42db-9fd0-f4782589ca18/volumes/kubernetes.io~projected/kube-api-access-wpgqz DeviceMajor:0 DeviceMinor:1469 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1573 DeviceMajor:0 DeviceMinor:1573 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/568e9bed-cbd2-49bd-84af-9e17f95a003f/volumes/kubernetes.io~projected/kube-api-access-flvlq DeviceMajor:0 DeviceMinor:1364 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/4c2975ec-e33d-4960-a708-277d41c79b15/volumes/kubernetes.io~secret/srv-cert DeviceMajor:0 DeviceMinor:1034 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/0de26264-46c2-4d29-97da-25a1681d6a8e/volumes/kubernetes.io~projected/kube-api-access-cdc7g DeviceMajor:0 DeviceMinor:1050 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/82464e92-4de2-48f0-8772-a489abb16898/volumes/kubernetes.io~projected/kube-api-access-4dk2b DeviceMajor:0 DeviceMinor:1052 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/5cd120c4a52499689b34a1d9baed806acbb93f12cd8a2bcbf5cdb4f6c23214df/userdata/shm DeviceMajor:0 DeviceMinor:1344 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1275 DeviceMajor:0 DeviceMinor:1275 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1186 DeviceMajor:0 DeviceMinor:1186 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8/volumes/kubernetes.io~projected/kube-api-access-jrdsv DeviceMajor:0 DeviceMinor:138 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/a722cda9-29a0-4b7f-8e1d-9a8950ed765a/volumes/kubernetes.io~secret/cluster-monitoring-operator-tls DeviceMajor:0 DeviceMinor:617 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-560 DeviceMajor:0 DeviceMinor:560 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/6010e9394b879952b424eb35b6227a92d6319988a7bfc5f908a8adc470e9463b/userdata/shm DeviceMajor:0 DeviceMinor:1077 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/48bd1d86-a6f2-439f-ab04-6a9a442bec42/volumes/kubernetes.io~secret/cert DeviceMajor:0 DeviceMinor:1036 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/efc2234f689663f4f73596bfe5f7a66235fb095a2337fa0c8bc412e1e08433fb/userdata/shm DeviceMajor:0 DeviceMinor:139 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1192 DeviceMajor:0 DeviceMinor:1192 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-366 DeviceMajor:0 DeviceMinor:366 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/21a9bd5a-3ade-48ef-8004-e0492daa85ba/volumes/kubernetes.io~projected/kube-api-access-r8bpw DeviceMajor:0 DeviceMinor:1476 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~secret/apiservice-cert DeviceMajor:0 DeviceMinor:524 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-110 DeviceMajor:0 DeviceMinor:110 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-371 DeviceMajor:0 DeviceMinor:371 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/aa4c28f3537eda89f3c42c0956c1461303fba100316bf76337e7c6ab1fb2e2e4/userdata/shm DeviceMajor:0 DeviceMinor:639 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c/volumes/kubernetes.io~projected/kube-api-access-ssssf DeviceMajor:0 DeviceMinor:153 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1268 DeviceMajor:0 DeviceMinor:1268 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1509 DeviceMajor:0 DeviceMinor:1509 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-167 DeviceMajor:0 DeviceMinor:167 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1480 DeviceMajor:0 DeviceMinor:1480 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/8114256db0301425275a52e5c6f4316fb7a7d5eed5146413b166f9ab3918f143/userdata/shm DeviceMajor:0 DeviceMinor:341 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-725 DeviceMajor:0 DeviceMinor:725 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/49051e6e-5a2f-45c8-bad0-374514a91c07/volumes/kubernetes.io~secret/cluster-olm-operator-serving-cert DeviceMajor:0 DeviceMinor:263 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1282 DeviceMajor:0 DeviceMinor:1282 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~projected/kube-api-access-ltjt6 DeviceMajor:0 DeviceMinor:811 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-972 DeviceMajor:0 DeviceMinor:972 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-997 DeviceMajor:0 DeviceMinor:997 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~secret/encryption-config DeviceMajor:0 DeviceMinor:561 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/a4fd453c-e667-4bdc-aa9e-3d95ff707200/volumes/kubernetes.io~projected/kube-api-access-kmhhw DeviceMajor:0 DeviceMinor:817 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1025 DeviceMajor:0 DeviceMinor:1025 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/2c3fab4d-05b8-42ec-8c91-91ff64562649/volumes/kubernetes.io~secret/apiservice-cert DeviceMajor:0 DeviceMinor:1039 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-62 DeviceMajor:0 DeviceMinor:62 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-205 DeviceMajor:0 DeviceMinor:205 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/a722cda9-29a0-4b7f-8e1d-9a8950ed765a/volumes/kubernetes.io~projected/kube-api-access-2lwgq DeviceMajor:0 DeviceMinor:276 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-483 DeviceMajor:0 DeviceMinor:483 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1184 DeviceMajor:0 DeviceMinor:1184 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1409 DeviceMajor:0 DeviceMinor:1409 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-97 DeviceMajor:0 DeviceMinor:97 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-213 DeviceMajor:0 DeviceMinor:213 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-902 DeviceMajor:0 DeviceMinor:902 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~projected/kube-api-access DeviceMajor:0 DeviceMinor:281 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/926263c4-ec5b-41cb-9c30-0c88f636035f/volumes/kubernetes.io~secret/marketplace-operator-metrics DeviceMajor:0 DeviceMinor:615 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/f12647ed460372c13fb857a5856ac53a5f123d196cc70548c9f47e2b35c490ce/userdata/shm DeviceMajor:0 DeviceMinor:1060 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~projected/kube-api-access DeviceMajor:0 DeviceMinor:288 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-511 DeviceMajor:0 DeviceMinor:511 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-60 DeviceMajor:0 DeviceMinor:60 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-588 DeviceMajor:0 DeviceMinor:588 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-398 DeviceMajor:0 DeviceMinor:398 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/b0fb01570845c0cbcb28b2232b0ade7593b0c0de030669cae5f3ef53544cda81/userdata/shm DeviceMajor:0 DeviceMinor:1056 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1583 DeviceMajor:0 DeviceMinor:1583 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/telemeter-client-tls DeviceMajor:0 DeviceMinor:783 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-695 DeviceMajor:0 DeviceMinor:695 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-786 DeviceMajor:0 DeviceMinor:786 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-850 DeviceMajor:0 DeviceMinor:850 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/45e53c8beb8eb2f205541b531720f581844942fc2ea4f6200401dc7c6d748d82/userdata/shm DeviceMajor:0 DeviceMinor:1312 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/ad7983dc1d046a27e0480393c3ce354f96d0968b079cf24fa1ebf196a469d66d/userdata/shm DeviceMajor:0 DeviceMinor:127 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-193 DeviceMajor:0 DeviceMinor:193 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/fc75409854b1370368a39cd91bcd5c4ac2aa319c2a8b9aaad1e7abab3fd56a49/userdata/shm DeviceMajor:0 DeviceMinor:179 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1020 DeviceMajor:0 DeviceMinor:1020 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/secret-telemeter-client DeviceMajor:0 DeviceMinor:785 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-618 DeviceMajor:0 DeviceMinor:618 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~projected/kube-api-access-twh85 DeviceMajor:0 DeviceMinor:299 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volume-subpaths/run-systemd/ovnkube-controller/6 DeviceMajor:0 DeviceMinor:24 Capacity:10102833152 Type:vfs Inodes:819200 HasInodes:true} {Device:/var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~secret/image-registry-operator-tls DeviceMajor:0 DeviceMinor:558 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/7f2b0f7328f1fdf55c43aa55c01436ada2488903a3cf2d212fd3c7469222fb7b/userdata/shm DeviceMajor:0 DeviceMinor:1024 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:810 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1195 DeviceMajor:0 DeviceMinor:1195 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/d22a14317ab6ff7dd171c6dc6d6a0e336ee5087a9085305232d6accf32502f02/userdata/shm DeviceMajor:0 DeviceMinor:54 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/b73ea430e90c45d058f4dc00766a5e44931feca970feb8976e3fca4580bc112f/userdata/shm DeviceMajor:0 DeviceMinor:95 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:255 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/58b95ae4-7f74-4777-8441-0a0ae28199e9/volumes/kubernetes.io~projected/kube-api-access-7fbtj DeviceMajor:0 DeviceMinor:1047 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/9dab351b1866858c81dfc4a1e5ed04a7dcaa8ee50017ec3f0b131de9c613ce86/userdata/shm DeviceMajor:0 DeviceMinor:1560 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-294 DeviceMajor:0 DeviceMinor:294 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-545 DeviceMajor:0 DeviceMinor:545 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1105 DeviceMajor:0 DeviceMinor:1105 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/4f32a485a101f0d25f8273536aa5f41f281533450b1891e0d6dbaa0dcf7b996e/userdata/shm DeviceMajor:0 DeviceMinor:1474 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/62f174ff-9f3c-4051-a34a-e55a59207171/volumes/kubernetes.io~secret/machine-approver-tls DeviceMajor:0 DeviceMinor:1144 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-513 DeviceMajor:0 DeviceMinor:513 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-434 DeviceMajor:0 DeviceMinor:434 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/58b95ae4-7f74-4777-8441-0a0ae28199e9/volumes/kubernetes.io~secret/cluster-storage-operator-serving-cert DeviceMajor:0 DeviceMinor:1033 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1099 DeviceMajor:0 DeviceMinor:1099 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:262 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-729 DeviceMajor:0 DeviceMinor:729 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-959 DeviceMajor:0 DeviceMinor:959 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1564 DeviceMajor:0 DeviceMinor:1564 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/2543decbaf3eadc4ac21f8d8b5c5c326c025c08f080ed2e80c25c76a6e2b34dd/userdata/shm DeviceMajor:0 DeviceMinor:1638 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1582 DeviceMajor:0 DeviceMinor:1582 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/11f563d5-89bb-433c-956a-6d5d2492e8f1/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:256 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-941 DeviceMajor:0 DeviceMinor:941 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1011 DeviceMajor:0 DeviceMinor:1011 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1208 DeviceMajor:0 DeviceMinor:1208 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd/volumes/kubernetes.io~projected/kube-api-access-vtsw9 DeviceMajor:0 DeviceMinor:1054 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e/volumes/kubernetes.io~secret/prometheus-operator-tls DeviceMajor:0 DeviceMinor:1445 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-474 DeviceMajor:0 DeviceMinor:474 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e/volumes/kubernetes.io~secret/prometheus-operator-kube-rbac-proxy-config DeviceMajor:0 DeviceMinor:1443 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-93 DeviceMajor:0 DeviceMinor:93 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-364 DeviceMajor:0 DeviceMinor:364 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/0d874913b26ce96decb20e6d175c667c355d6a65334e625b074f3811b11e3027/userdata/shm DeviceMajor:0 DeviceMinor:541 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-693 DeviceMajor:0 DeviceMinor:693 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-652 DeviceMajor:0 DeviceMinor:652 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-387 DeviceMajor:0 DeviceMinor:387 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-744 DeviceMajor:0 DeviceMinor:744 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/b7508cd3-7421-4ab9-be1f-318db2853596/volumes/kubernetes.io~secret/cert DeviceMajor:0 DeviceMinor:645 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/4446e01522dc26b79f498f8e0dc137457645ab081cac6529b44e9b557b9d72f8/userdata/shm DeviceMajor:0 DeviceMinor:136 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-354 DeviceMajor:0 DeviceMinor:354 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/2421d54e72b15809966ab160149e8662f98357fd3da518130cc175e2f9f3791a/userdata/shm DeviceMajor:0 DeviceMinor:1446 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1644 DeviceMajor:0 DeviceMinor:1644 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1434 DeviceMajor:0 DeviceMinor:1434 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-291 DeviceMajor:0 DeviceMinor:291 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/6728a458-098b-44f9-8c8f-b0d76c5825fa/volumes/kubernetes.io~secret/proxy-tls DeviceMajor:0 DeviceMinor:1310 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/2c3fab4d-05b8-42ec-8c91-91ff64562649/volumes/kubernetes.io~projected/kube-api-access-l8wnp DeviceMajor:0 DeviceMinor:1051 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-995 DeviceMajor:0 DeviceMinor:995 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-129 DeviceMajor:0 DeviceMinor:129 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/f7a08359-0379-4364-8b0c-ddb58ff605f4/volumes/kubernetes.io~secret/etcd-client DeviceMajor:0 DeviceMinor:261 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-356 DeviceMajor:0 DeviceMinor:356 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1008 DeviceMajor:0 DeviceMinor:1008 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-197 DeviceMajor:0 DeviceMinor:197 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-650 DeviceMajor:0 DeviceMinor:650 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/d4750111f1d0e54879f82e850c1814dfc68424e9b8c03f07177ca16103b8a79d/userdata/shm DeviceMajor:0 DeviceMinor:1066 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1321 DeviceMajor:0 DeviceMinor:1321 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-151 DeviceMajor:0 DeviceMinor:151 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-566 DeviceMajor:0 DeviceMinor:566 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/568e9bed-cbd2-49bd-84af-9e17f95a003f/volumes/kubernetes.io~secret/node-bootstrap-token DeviceMajor:0 DeviceMinor:1363 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1507 DeviceMajor:0 DeviceMinor:1507 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~secret/profile-collector-cert DeviceMajor:0 DeviceMinor:254 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c/volumes/kubernetes.io~secret/metrics-certs DeviceMajor:0 DeviceMinor:614 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:733 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-100 DeviceMajor:0 DeviceMinor:100 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-131 DeviceMajor:0 DeviceMinor:131 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-133 DeviceMajor:0 DeviceMinor:133 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~projected/kube-api-access DeviceMajor:0 DeviceMinor:290 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~projected/kube-api-access-m88gw DeviceMajor:0 DeviceMinor:611 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-195 DeviceMajor:0 DeviceMinor:195 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/592373ae-a952-4dd3-a9bc-f9c9c19c0802/volumes/kubernetes.io~projected/ca-certs DeviceMajor:0 DeviceMinor:819 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-57 DeviceMajor:0 DeviceMinor:57 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1407 DeviceMajor:0 DeviceMinor:1407 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/24444474-aa9a-4a0f-8b4d-90f0009e0dc7/volumes/kubernetes.io~projected/kube-api-access-lrwt6 DeviceMajor:0 DeviceMinor:690 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1287 DeviceMajor:0 DeviceMinor:1287 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-156 DeviceMajor:0 DeviceMinor:156 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1270 DeviceMajor:0 DeviceMinor:1270 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1621 DeviceMajor:0 DeviceMinor:1621 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-209 DeviceMajor:0 DeviceMinor:209 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/5b905a3a36c591b5f3704d14e85aafe47726c309b50985c9308b5ed0d4a90267/userdata/shm DeviceMajor:0 DeviceMinor:346 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~secret/metrics-tls DeviceMajor:0 DeviceMinor:556 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/07cb2a5b59f4a1cf9ae60c0a62bf4e44ad34c98cbd92efe399d17d750ec292c6/userdata/shm DeviceMajor:0 DeviceMinor:1176 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-390 DeviceMajor:0 DeviceMinor:390 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1292 DeviceMajor:0 DeviceMinor:1292 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-496 DeviceMajor:0 DeviceMinor:496 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1600 DeviceMajor:0 DeviceMinor:1600 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/1e69ce9e-4e6f-4015-9ba6-5a7942570190/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:257 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/77b83e99ece1e07058ac092ebb18d76d85bb5a90292440cd154e967f63253776/userdata/shm DeviceMajor:0 DeviceMinor:626 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/80b1488f509c166b5b116f44dccb9102304f263ee75bb4a43991cef27c102c84/userdata/shm DeviceMajor:0 DeviceMinor:532 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-653 DeviceMajor:0 DeviceMinor:653 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/444f8808-e454-4015-9e20-429e715a08c7/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:260 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/c22d947f-a5b6-4f24-b142-dd201c46293b/volumes/kubernetes.io~projected/kube-api-access-98qt8 DeviceMajor:0 DeviceMinor:273 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-749 DeviceMajor:0 DeviceMinor:749 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1119 DeviceMajor:0 DeviceMinor:1119 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-470 DeviceMajor:0 DeviceMinor:470 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/8c649a16-c187-412e-b5da-62a00bee38ab/volumes/kubernetes.io~secret/package-server-manager-serving-cert DeviceMajor:0 DeviceMinor:613 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:1038 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/0babac16-164a-405a-a86c-30524118f046/volumes/kubernetes.io~secret/tls-certificates DeviceMajor:0 DeviceMinor:1337 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/da9d48c9-6346-4c9f-a690-f7419499c3e6/volumes/kubernetes.io~projected/kube-api-access-7824l DeviceMajor:0 DeviceMinor:1339 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-604 DeviceMajor:0 DeviceMinor:604 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/2c8d17056f01fe2b63601e7d3f2b689f6a1b5945d124ed1dd3ebcdd26090636d/userdata/shm DeviceMajor:0 DeviceMinor:534 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/01488ac4-313d-48d5-9e86-7b2011b9e91e/volumes/kubernetes.io~projected/kube-api-access-2xjx8 DeviceMajor:0 DeviceMinor:1048 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/af7078c2-ad4f-415b-ba2f-77f5145c3b3f/volumes/kubernetes.io~secret/cloud-controller-manager-operator-tls DeviceMajor:0 DeviceMinor:1040 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/8904f4db7f08be6823cd5712612b6b29bb4d4506ff20a7f693a2b99016d20918/userdata/shm DeviceMajor:0 DeviceMinor:1177 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/54f9f9d4-8c31-4636-8010-22ee11b9b323/volumes/kubernetes.io~projected/kube-api-access-sv2tg DeviceMajor:0 DeviceMinor:1475 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/6ead10efe7a643d9b0dd883cc19f8ef852a5658bc79c235b7c7c6b5de2e97811/userdata/shm DeviceMajor:0 DeviceMinor:189 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-228 DeviceMajor:0 DeviceMinor:228 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-406 DeviceMajor:0 DeviceMinor:406 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/8066239534fa93e4ea7b428954f42f9cdca327545a8ae0d3c647f8ee2fd24e03/userdata/shm DeviceMajor:0 DeviceMinor:531 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/48bd1d86-a6f2-439f-ab04-6a9a442bec42/volumes/kubernetes.io~projected/kube-api-access-f69hj DeviceMajor:0 DeviceMinor:1045 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/897152fdb47d25f0a5016944193be33416b9e61d9cbcab904f0e8c6b63de2a36/userdata/shm DeviceMajor:0 DeviceMinor:1062 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1097 DeviceMajor:0 DeviceMinor:1097 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1190 DeviceMajor:0 DeviceMinor:1190 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~secret/node-tuning-operator-tls DeviceMajor:0 DeviceMinor:523 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-641 DeviceMajor:0 DeviceMinor:641 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-907 DeviceMajor:0 DeviceMinor:907 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-559 DeviceMajor:0 DeviceMinor:559 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/5be78d3569389b67ef6d6c6575e7c97a75617917f6e7f31b9b05a692b9e12292/userdata/shm DeviceMajor:0 DeviceMinor:55 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/33616deb-ea10-4a38-8681-ab023b526b11/volumes/kubernetes.io~empty-dir/tmp DeviceMajor:0 DeviceMinor:580 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-711 DeviceMajor:0 DeviceMinor:711 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/a4fd453c-e667-4bdc-aa9e-3d95ff707200/volumes/kubernetes.io~projected/ca-certs DeviceMajor:0 DeviceMinor:816 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:10102833152 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~secret/ovn-control-plane-metrics-cert DeviceMajor:0 DeviceMinor:163 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/b760849c-8d83-47da-8677-68445c143bef/volumes/kubernetes.io~projected/kube-api-access-jfz87 DeviceMajor:0 DeviceMinor:164 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/3332b604-6f0d-4243-bce1-b3ab76e58c78/volumes/kubernetes.io~projected/kube-api-access-6gfgm DeviceMajor:0 DeviceMinor:1637 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-606 DeviceMajor:0 DeviceMinor:606 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1448 DeviceMajor:0 DeviceMinor:1448 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1399 DeviceMajor:0 DeviceMinor:1399 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1595 DeviceMajor:0 DeviceMinor:1595 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/21a9bd5a-3ade-48ef-8004-e0492daa85ba/volumes/kubernetes.io~secret/node-exporter-kube-rbac-proxy-config DeviceMajor:0 DeviceMinor:1471 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/3332b604-6f0d-4243-bce1-b3ab76e58c78/volumes/kubernetes.io~secret/webhook-certs DeviceMajor:0 DeviceMinor:1636 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/bd5095616b6a8494a79971c731d5ef49c3bae5cbd4111190f1e5dfde0e3d9554/userdata/shm DeviceMajor:0 DeviceMinor:812 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-824 DeviceMajor:0 DeviceMinor:824 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/24444474-aa9a-4a0f-8b4d-90f0009e0dc7/volumes/kubernetes.io~secret/control-plane-machine-set-operator-tls DeviceMajor:0 DeviceMinor:689 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-72 DeviceMajor:0 DeviceMinor:72 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~secret/webhook-cert DeviceMajor:0 DeviceMinor:187 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-697 DeviceMajor:0 DeviceMinor:697 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/0de26264-46c2-4d29-97da-25a1681d6a8e/volumes/kubernetes.io~secret/samples-operator-tls DeviceMajor:0 DeviceMinor:1037 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1314 DeviceMajor:0 DeviceMinor:1314 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~projected/kube-api-access-hskm2 DeviceMajor:0 DeviceMinor:1341 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/b7508cd3-7421-4ab9-be1f-318db2853596/volumes/kubernetes.io~projected/kube-api-access-kgvv6 DeviceMajor:0 DeviceMinor:644 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-377 DeviceMajor:0 DeviceMinor:377 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-392 DeviceMajor:0 DeviceMinor:392 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-867 Dec 05 10:47:20.118914 master-0 kubenswrapper[24928]: DeviceMajor:0 DeviceMinor:867 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/9e8deb0a79a04a088af836af85742626a8778be7f61d22448e3a0b846ff8b05a/userdata/shm DeviceMajor:0 DeviceMinor:1004 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/5a1bdc70-6412-47e0-8330-04d796cc8d55/volumes/kubernetes.io~projected/kube-api-access-5dtfn DeviceMajor:0 DeviceMinor:292 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-410 DeviceMajor:0 DeviceMinor:410 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/61b925377e25c3f9cf64e2a949ea1fe86f2ea5c4e951a4e7adc2aa3b0cb3336d/userdata/shm DeviceMajor:0 DeviceMinor:1343 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-939 DeviceMajor:0 DeviceMinor:939 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-77 DeviceMajor:0 DeviceMinor:77 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/secret-telemeter-client-kube-rbac-proxy-config DeviceMajor:0 DeviceMinor:784 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:overlay_0-154 DeviceMajor:0 DeviceMinor:154 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/d1c3b7dd-f25e-4983-8a94-084f863fd5b9/volumes/kubernetes.io~projected/kube-api-access-fkmvj DeviceMajor:0 DeviceMinor:274 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/1fd0a349-0f66-4c85-95e4-8bbc96648c05/volumes/kubernetes.io~projected/kube-api-access DeviceMajor:0 DeviceMinor:884 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-379 DeviceMajor:0 DeviceMinor:379 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1641 DeviceMajor:0 DeviceMinor:1641 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-792 DeviceMajor:0 DeviceMinor:792 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1505 DeviceMajor:0 DeviceMinor:1505 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-547 DeviceMajor:0 DeviceMinor:547 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/4c2975ec-e33d-4960-a708-277d41c79b15/volumes/kubernetes.io~secret/profile-collector-cert DeviceMajor:0 DeviceMinor:1031 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/e88afc2116e69a8053c275a20f5cc4b6af9c5241405fc2bf27810f224b5549fb/userdata/shm DeviceMajor:0 DeviceMinor:1146 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-199 DeviceMajor:0 DeviceMinor:199 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-385 DeviceMajor:0 DeviceMinor:385 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-400 DeviceMajor:0 DeviceMinor:400 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a/volumes/kubernetes.io~projected/kube-api-access-pdsh9 DeviceMajor:0 DeviceMinor:472 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-147 DeviceMajor:0 DeviceMinor:147 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/1bde7edcbb5c913afcdfed1e9b165ffe883b909863e89e771b54560c18c0ac2a/userdata/shm DeviceMajor:0 DeviceMinor:954 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/82464e92-4de2-48f0-8772-a489abb16898/volumes/kubernetes.io~secret/machine-api-operator-tls DeviceMajor:0 DeviceMinor:1032 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1353 DeviceMajor:0 DeviceMinor:1353 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/70e7f53d39e81bfad8819ec3c2244577ab77bdd0b9a5512ba3a7cb61ae8bafb4/userdata/shm DeviceMajor:0 DeviceMinor:473 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-848 DeviceMajor:0 DeviceMinor:848 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:832 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~secret/default-certificate DeviceMajor:0 DeviceMinor:1335 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1101 DeviceMajor:0 DeviceMinor:1101 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/volumes/kubernetes.io~projected/kube-api-access-wg9zq DeviceMajor:0 DeviceMinor:188 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/eb290494-a456-4f0e-9afc-f20abab1a1bf/volumes/kubernetes.io~projected/kube-api-access-qn7rj DeviceMajor:0 DeviceMinor:267 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-943 DeviceMajor:0 DeviceMinor:943 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-438 DeviceMajor:0 DeviceMinor:438 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-527 DeviceMajor:0 DeviceMinor:527 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/48bd1d86-a6f2-439f-ab04-6a9a442bec42/volumes/kubernetes.io~secret/cluster-baremetal-operator-tls DeviceMajor:0 DeviceMinor:1041 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~secret/secret-metrics-server-tls DeviceMajor:0 DeviceMinor:1556 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/fd58232c-a81a-4aee-8b2c-5ffcdded2e23/volumes/kubernetes.io~projected/kube-api-access-h2vd4 DeviceMajor:0 DeviceMinor:325 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/33616deb-ea10-4a38-8681-ab023b526b11/volumes/kubernetes.io~projected/kube-api-access-rsrkp DeviceMajor:0 DeviceMinor:582 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-846 DeviceMajor:0 DeviceMinor:846 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/0f19c2361e4856d55a1f7cfcf2bbf2e6aae6be0b4b861e0e68ba7772ccbc2487/userdata/shm DeviceMajor:0 DeviceMinor:875 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-568 DeviceMajor:0 DeviceMinor:568 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/487c60a3656ecce96053a1bf83cb0f4b361d27909df0d101db089d62b7524b7b/userdata/shm DeviceMajor:0 DeviceMinor:619 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/445d75af-d072-4fa0-91a7-f3fa579b9ca9/volumes/kubernetes.io~secret/stats-auth DeviceMajor:0 DeviceMinor:1338 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1264 DeviceMajor:0 DeviceMinor:1264 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~secret/federate-client-tls DeviceMajor:0 DeviceMinor:782 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/12d79e8d025da0e5c89728d8b35230c8c9764259d39417d2ff38f00dc9c263f6/userdata/shm DeviceMajor:0 DeviceMinor:349 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/088f93fb71215e03078ad676d01b6c7b68c855017eff77ac0edbcdeaa645fd6c/userdata/shm DeviceMajor:0 DeviceMinor:503 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/0e8ac47ff99876f6c126e479233a75b5de6a9aabecbf3b414116156cf0e23f86/userdata/shm DeviceMajor:0 DeviceMinor:894 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/97d00d6aa4d6661e2be349a4e20636f067fc69912c7110a63d99e820d5970ef4/userdata/shm DeviceMajor:0 DeviceMinor:718 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1532 DeviceMajor:0 DeviceMinor:1532 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-789 DeviceMajor:0 DeviceMinor:789 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-360 DeviceMajor:0 DeviceMinor:360 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1123 DeviceMajor:0 DeviceMinor:1123 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/52b24a28414c04f2574070f038561f387ec44118f13f6cac5eab1761c0750f3a/userdata/shm DeviceMajor:0 DeviceMinor:1173 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/af7078c2-ad4f-415b-ba2f-77f5145c3b3f/volumes/kubernetes.io~projected/kube-api-access-ljdt5 DeviceMajor:0 DeviceMinor:1055 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-404 DeviceMajor:0 DeviceMinor:404 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/97c0ac3e4d477fff0dd9953d178193fc87456dca36efdf5e88346f7b00aeff5b/userdata/shm DeviceMajor:0 DeviceMinor:525 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/a669846f3b9be1321aadc153fc634864e5c4e6b7c212feee982c065424fad087/userdata/shm DeviceMajor:0 DeviceMinor:533 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-861 DeviceMajor:0 DeviceMinor:861 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-102 DeviceMajor:0 DeviceMinor:102 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/5a1bdc70-6412-47e0-8330-04d796cc8d55/volumes/kubernetes.io~secret/metrics-tls DeviceMajor:0 DeviceMinor:557 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-691 DeviceMajor:0 DeviceMinor:691 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-747 DeviceMajor:0 DeviceMinor:747 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1202 DeviceMajor:0 DeviceMinor:1202 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-637 DeviceMajor:0 DeviceMinor:637 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-863 DeviceMajor:0 DeviceMinor:863 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-203 DeviceMajor:0 DeviceMinor:203 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1121 DeviceMajor:0 DeviceMinor:1121 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/0aa6e138-4b1f-4047-8255-a2b14d044588/volumes/kubernetes.io~projected/kube-api-access-wnzgs DeviceMajor:0 DeviceMinor:1168 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/568e9bed-cbd2-49bd-84af-9e17f95a003f/volumes/kubernetes.io~secret/certs DeviceMajor:0 DeviceMinor:1362 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-90 DeviceMajor:0 DeviceMinor:90 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/384f1c47-f511-4b58-aa7f-71aef6ef91a9/volumes/kubernetes.io~projected/kube-api-access-lgtnt DeviceMajor:0 DeviceMinor:1022 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-76 DeviceMajor:0 DeviceMinor:76 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/fedfe3c8bf160c89d4d691575ab31b377473c05a2dd753786d2ef1a80ea5d562/userdata/shm DeviceMajor:0 DeviceMinor:347 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/volumes/kubernetes.io~projected/kube-api-access-ml74w DeviceMajor:0 DeviceMinor:465 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-104 DeviceMajor:0 DeviceMinor:104 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/62f174ff-9f3c-4051-a34a-e55a59207171/volumes/kubernetes.io~projected/kube-api-access-r5p8s DeviceMajor:0 DeviceMinor:1145 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1347 DeviceMajor:0 DeviceMinor:1347 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1577 DeviceMajor:0 DeviceMinor:1577 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-358 DeviceMajor:0 DeviceMinor:358 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/59a797eb008b1a85db4180d24d857e8f08956acdd19cbd4d960f8adc3117a02b/userdata/shm DeviceMajor:0 DeviceMinor:555 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/7c5339e633ef9bd66b822a691a1a94b12e50803e5297f66f5662bfb59a1a2a6c/userdata/shm DeviceMajor:0 DeviceMinor:1072 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1125 DeviceMajor:0 DeviceMinor:1125 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/f7b29f89-e42d-4e53-ad14-05efdce933f0/volumes/kubernetes.io~projected/kube-api-access-khfxz DeviceMajor:0 DeviceMinor:296 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/926263c4-ec5b-41cb-9c30-0c88f636035f/volumes/kubernetes.io~projected/kube-api-access-fncz7 DeviceMajor:0 DeviceMinor:318 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/ee0b4a7f-1b96-4304-bef0-fb575e0e431c/volumes/kubernetes.io~secret/cloud-credential-operator-serving-cert DeviceMajor:0 DeviceMinor:1042 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1058 DeviceMajor:0 DeviceMinor:1058 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/22676fac-b770-4937-9bee-7478bd1babb7/volumes/kubernetes.io~projected/kube-api-access-vkqz7 DeviceMajor:0 DeviceMinor:287 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-570 DeviceMajor:0 DeviceMinor:570 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-981 DeviceMajor:0 DeviceMinor:981 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/54f9f9d4-8c31-4636-8010-22ee11b9b323/volumes/kubernetes.io~secret/kube-state-metrics-kube-rbac-proxy-config DeviceMajor:0 DeviceMinor:1473 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-316 DeviceMajor:0 DeviceMinor:316 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/720a7467-ce93-4d48-82ec-9ad0922d99c2/volumes/kubernetes.io~projected/kube-api-access-5n7tf DeviceMajor:0 DeviceMinor:482 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1150 DeviceMajor:0 DeviceMinor:1150 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/6a9f011f-36f1-4308-a365-69425c186c7f/volumes/kubernetes.io~projected/kube-api-access-mvnxf DeviceMajor:0 DeviceMinor:137 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/f77a8c44878083fb1f5b837847555bd296fce89d49bff4c5ba5fc4a9953922c2/userdata/shm DeviceMajor:0 DeviceMinor:1073 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1367 DeviceMajor:0 DeviceMinor:1367 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-181 DeviceMajor:0 DeviceMinor:181 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/0e02576ddbd856f2a1c30d4bbdcb0b27384c73dfe012be69bbb5c2168f8df7fd/userdata/shm DeviceMajor:0 DeviceMinor:621 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1351 DeviceMajor:0 DeviceMinor:1351 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-999 DeviceMajor:0 DeviceMinor:999 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1484 DeviceMajor:0 DeviceMinor:1484 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1576 DeviceMajor:0 DeviceMinor:1576 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~projected/kube-api-access-tjgc4 DeviceMajor:0 DeviceMinor:172 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/b44ac27d1b7f531e11c3f4d352d89c4a7679a795ec64417e43c87c5d1e725e90/userdata/shm DeviceMajor:0 DeviceMinor:314 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-586 DeviceMajor:0 DeviceMinor:586 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-909 DeviceMajor:0 DeviceMinor:909 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1376 DeviceMajor:0 DeviceMinor:1376 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/d5198438-06ae-4e63-a7e3-950ba23bba9c/volumes/kubernetes.io~secret/etcd-client DeviceMajor:0 DeviceMinor:562 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082/volumes/kubernetes.io~secret/srv-cert DeviceMajor:0 DeviceMinor:616 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-913 DeviceMajor:0 DeviceMinor:913 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1486 DeviceMajor:0 DeviceMinor:1486 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/54f9f9d4-8c31-4636-8010-22ee11b9b323/volumes/kubernetes.io~secret/kube-state-metrics-tls DeviceMajor:0 DeviceMinor:1472 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-881 DeviceMajor:0 DeviceMinor:881 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1103 DeviceMajor:0 DeviceMinor:1103 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-183 DeviceMajor:0 DeviceMinor:183 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/283122ba-be1c-4516-bd0f-df41c13c098b/volumes/kubernetes.io~projected/kube-api-access-vhjhk DeviceMajor:0 DeviceMinor:275 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-674 DeviceMajor:0 DeviceMinor:674 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~secret/encryption-config DeviceMajor:0 DeviceMinor:809 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1266 DeviceMajor:0 DeviceMinor:1266 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/bdfbd9daf37e036f9b2d20314905a9aa48f68e21954c4e133ee980fe10b0d8d2/userdata/shm DeviceMajor:0 DeviceMinor:343 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/33616deb-ea10-4a38-8681-ab023b526b11/volumes/kubernetes.io~empty-dir/etc-tuned DeviceMajor:0 DeviceMinor:581 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0/volumes/kubernetes.io~projected/kube-api-access-95jmp DeviceMajor:0 DeviceMinor:860 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1188 DeviceMajor:0 DeviceMinor:1188 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1671 DeviceMajor:0 DeviceMinor:1671 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-381 DeviceMajor:0 DeviceMinor:381 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-487 DeviceMajor:0 DeviceMinor:487 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-389 DeviceMajor:0 DeviceMinor:389 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-432 DeviceMajor:0 DeviceMinor:432 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/6607da6ec4214a3fe2d315e70d9756b15e987231bb8310517d849da88c8e7b15/userdata/shm DeviceMajor:0 DeviceMinor:165 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-322 DeviceMajor:0 DeviceMinor:322 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-499 DeviceMajor:0 DeviceMinor:499 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/1f132702-304b-46ff-b428-6df1deeffec3/volumes/kubernetes.io~projected/kube-api-access-vkzkh DeviceMajor:0 DeviceMinor:1049 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/812553092d13019021bace1ffb05fbafd76c83d356691f1b938f38559234f787/userdata/shm DeviceMajor:0 DeviceMinor:1340 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/d9ff1ae2-8547-42db-9fd0-f4782589ca18/volumes/kubernetes.io~secret/openshift-state-metrics-tls DeviceMajor:0 DeviceMinor:1467 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-631 DeviceMajor:0 DeviceMinor:631 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-120 DeviceMajor:0 DeviceMinor:120 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-158 DeviceMajor:0 DeviceMinor:158 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/89c8e517113e00942d58576e87acb9fd6a3e8f726e30d612913a3d827c50d10f/userdata/shm DeviceMajor:0 DeviceMinor:485 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/baa812e4-b011-41ee-82ab-8f571f0b7e0a/volumes/kubernetes.io~projected/kube-api-access-28qvt DeviceMajor:0 DeviceMinor:1167 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/4825316a-ea9f-4d3d-838b-fa809a6e49c7/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:258 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/c93da22a302a383f4fdc18e25cf772f0d9d551ca131191ed837accfc6f610c6f/userdata/shm DeviceMajor:0 DeviceMinor:1064 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-723 DeviceMajor:0 DeviceMinor:723 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-790 DeviceMajor:0 DeviceMinor:790 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-394 DeviceMajor:0 DeviceMinor:394 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/c22e02aa77c0fa78cb950ff2c4f24a8244e5ac70b15a797f17b5f262893f0220/userdata/shm DeviceMajor:0 DeviceMinor:1068 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1459 DeviceMajor:0 DeviceMinor:1459 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-92 DeviceMajor:0 DeviceMinor:92 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e/volumes/kubernetes.io~projected/kube-api-access-g88bl DeviceMajor:0 DeviceMinor:1444 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1391 DeviceMajor:0 DeviceMinor:1391 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/4833f667ac4451a150ed477789be72c08b65162d8c7bbf29e44cd656299dbd63/userdata/shm DeviceMajor:0 DeviceMinor:375 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/62e34bf231128449a630a3dad98495dd55715547e65586a815afe200566d05a0/userdata/shm DeviceMajor:0 DeviceMinor:622 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:/run/containers/storage/overlay-containers/35b4f066f81dc683769e8a58a956f46b65324a32b3270c15056ef67e541a8fdd/userdata/shm DeviceMajor:0 DeviceMinor:1175 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-83 DeviceMajor:0 DeviceMinor:83 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-185 DeviceMajor:0 DeviceMinor:185 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-731 DeviceMajor:0 DeviceMinor:731 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/58baad85-de54-49e7-a13e-d470d9c50d11/volumes/kubernetes.io~secret/metrics-tls DeviceMajor:0 DeviceMinor:739 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/87909f47-f2d7-46f8-a1c8-27336cdcce5d/volumes/kubernetes.io~projected/kube-api-access-rjrgm DeviceMajor:0 DeviceMinor:323 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/ca45c52e-fb30-4e7c-8c3f-e685c0909916/volumes/kubernetes.io~secret/etcd-client DeviceMajor:0 DeviceMinor:808 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-961 DeviceMajor:0 DeviceMinor:961 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-141 DeviceMajor:0 DeviceMinor:141 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~secret/client-ca-bundle DeviceMajor:0 DeviceMinor:1558 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1578 DeviceMajor:0 DeviceMinor:1578 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-44 DeviceMajor:0 DeviceMinor:44 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/495ba1ea-f844-43ec-8be7-47e738f5428a/volumes/kubernetes.io~secret/ovn-node-metrics-cert DeviceMajor:0 DeviceMinor:171 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-402 DeviceMajor:0 DeviceMinor:402 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-877 DeviceMajor:0 DeviceMinor:877 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1655 DeviceMajor:0 DeviceMinor:1655 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-191 DeviceMajor:0 DeviceMinor:191 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-268 DeviceMajor:0 DeviceMinor:268 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1316 DeviceMajor:0 DeviceMinor:1316 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes/kubernetes.io~secret/secret-metrics-client-certs DeviceMajor:0 DeviceMinor:1557 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-236 DeviceMajor:0 DeviceMinor:236 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/aa871dfa794fea5d166618a43475fe935ef7d4d3799afbc3d92debffe4b76081/userdata/shm DeviceMajor:0 DeviceMinor:623 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1349 DeviceMajor:0 DeviceMinor:1349 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/1f132702-304b-46ff-b428-6df1deeffec3/volumes/kubernetes.io~secret/cert DeviceMajor:0 DeviceMinor:1043 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1204 DeviceMajor:0 DeviceMinor:1204 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-1646 DeviceMajor:0 DeviceMinor:1646 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-422 DeviceMajor:0 DeviceMinor:422 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/489b6771-aa16-469f-9883-9a9233d3c379/volumes/kubernetes.io~projected/kube-api-access-9pvrn DeviceMajor:0 DeviceMinor:1166 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-112 DeviceMajor:0 DeviceMinor:112 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes.io~secret/serving-cert DeviceMajor:0 DeviceMinor:264 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-529 DeviceMajor:0 DeviceMinor:529 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-945 DeviceMajor:0 DeviceMinor:945 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-655 DeviceMajor:0 DeviceMinor:655 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:overlay_0-633 DeviceMajor:0 DeviceMinor:633 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/b7e7f216-f9af-41df-a0b0-df2a76b9b72a/volumes/kubernetes.io~projected/kube-api-access-rvw27 DeviceMajor:0 DeviceMinor:1170 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:/var/lib/kubelet/pods/d95a56ba-c940-4e3e-aed6-d8c04f1871b6/volumes/kubernetes.io~projected/kube-api-access-2fbss DeviceMajor:0 DeviceMinor:313 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-408 DeviceMajor:0 DeviceMinor:408 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/run/containers/storage/overlay-containers/eecf6d3a967ce077414ad07bec5a21f5acf63791b8f14de4d8d34c6cfe91444a/userdata/shm DeviceMajor:0 DeviceMinor:468 Capacity:67108864 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-601 DeviceMajor:0 DeviceMinor:601 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/6f76d12f-5406-47e2-8337-2f50e35376d6/volumes/kubernetes.io~projected/kube-api-access-fzwfq DeviceMajor:0 DeviceMinor:266 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1294 DeviceMajor:0 DeviceMinor:1294 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/97192d4d-15d3-4740-82f3-d0d45f9fe7b9/volumes/kubernetes.io~projected/kube-api-access-nn6sc DeviceMajor:0 DeviceMinor:500 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-705 DeviceMajor:0 DeviceMinor:705 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/592373ae-a952-4dd3-a9bc-f9c9c19c0802/volumes/kubernetes.io~projected/kube-api-access-rd87p DeviceMajor:0 DeviceMinor:818 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-990 DeviceMajor:0 DeviceMinor:990 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:25257078784 Type:vfs Inodes:1048576 HasInodes:true} {Device:overlay_0-826 DeviceMajor:0 DeviceMinor:826 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/2c3fab4d-05b8-42ec-8c91-91ff64562649/volumes/kubernetes.io~secret/webhook-cert DeviceMajor:0 DeviceMinor:1035 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1534 DeviceMajor:0 DeviceMinor:1534 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/58baad85-de54-49e7-a13e-d470d9c50d11/volumes/kubernetes.io~projected/kube-api-access-jx7jr DeviceMajor:0 DeviceMinor:722 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-1206 DeviceMajor:0 DeviceMinor:1206 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true} {Device:/var/lib/kubelet/pods/d9ff1ae2-8547-42db-9fd0-f4782589ca18/volumes/kubernetes.io~secret/openshift-state-metrics-kube-rbac-proxy-config DeviceMajor:0 DeviceMinor:1468 Capacity:49335554048 Type:vfs Inodes:6166278 HasInodes:true} {Device:overlay_0-803 DeviceMajor:0 DeviceMinor:803 Capacity:214143315968 Type:vfs Inodes:104594880 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none} 252:16:{Name:vdb Major:252 Minor:16 Size:21474836480 Scheduler:none} 252:32:{Name:vdc Major:252 Minor:32 Size:21474836480 Scheduler:none} 252:48:{Name:vdd Major:252 Minor:48 Size:21474836480 Scheduler:none} 252:64:{Name:vde Major:252 Minor:64 Size:21474836480 Scheduler:none}] NetworkDevices:[{Name:088f93fb71215e0 MacAddress:d6:d8:c8:b2:84:31 Speed:10000 Mtu:8900} {Name:0d2211dcab91af9 MacAddress:ce:f0:13:83:60:f8 Speed:10000 Mtu:8900} {Name:0d874913b26ce96 MacAddress:c2:2c:78:76:b8:91 Speed:10000 Mtu:8900} {Name:0e02576ddbd856f MacAddress:b6:57:9d:a0:50:c2 Speed:10000 Mtu:8900} {Name:0f19c2361e4856d MacAddress:de:9d:6b:1f:c6:22 Speed:10000 Mtu:8900} {Name:12d79e8d025da0e MacAddress:b2:de:cd:56:04:ba Speed:10000 Mtu:8900} {Name:1bde7edcbb5c913 MacAddress:96:4a:e4:8d:40:c2 Speed:10000 Mtu:8900} {Name:2421d54e72b1580 MacAddress:c6:73:58:aa:87:26 Speed:10000 Mtu:8900} {Name:2543decbaf3eadc MacAddress:a2:27:b5:52:c2:67 Speed:10000 Mtu:8900} {Name:2c8d17056f01fe2 MacAddress:16:68:f3:64:e8:68 Speed:10000 Mtu:8900} {Name:2cf0d485e10f74f MacAddress:42:87:c7:2c:be:ed Speed:10000 Mtu:8900} {Name:35b4f066f81dc68 MacAddress:9a:ca:2e:36:56:97 Speed:10000 Mtu:8900} {Name:45e53c8beb8eb2f MacAddress:9e:d1:bf:54:2b:b1 Speed:10000 Mtu:8900} {Name:4833f667ac4451a MacAddress:0e:d6:8b:f6:6f:9f Speed:10000 Mtu:8900} {Name:487c60a3656ecce MacAddress:7a:4f:a0:fa:d1:ef Speed:10000 Mtu:8900} {Name:4c76b1c660e01cc MacAddress:6e:d6:64:9d:d2:f6 Speed:10000 Mtu:8900} {Name:4f32a485a101f0d MacAddress:ca:3d:55:65:58:96 Speed:10000 Mtu:8900} {Name:5046caa4441a176 MacAddress:02:7c:a4:7b:2f:ed Speed:10000 Mtu:8900} {Name:52b24a28414c04f MacAddress:2e:7b:ca:05:64:85 Speed:10000 Mtu:8900} {Name:59a797eb008b1a8 MacAddress:1e:9b:b3:d2:66:af Speed:10000 Mtu:8900} {Name:5ab705ed0c0dfea MacAddress:66:13:58:48:4d:0c Speed:10000 Mtu:8900} {Name:5b905a3a36c591b MacAddress:4a:f7:24:e5:b8:21 Speed:10000 Mtu:8900} {Name:6010e9394b87995 MacAddress:d6:11:a1:82:b3:a7 Speed:10000 Mtu:8900} {Name:61b925377e25c3f MacAddress:92:ee:5b:83:c8:1c Speed:10000 Mtu:8900} {Name:62e34bf23112844 MacAddress:76:d9:db:85:2e:2d Speed:10000 Mtu:8900} {Name:70e7f53d39e81bf MacAddress:fe:c7:bc:67:b8:7b Speed:10000 Mtu:8900} {Name:77b83e99ece1e07 MacAddress:9a:c7:8f:85:db:37 Speed:10000 Mtu:8900} {Name:7c5339e633ef9bd MacAddress:ce:fa:a9:33:28:73 Speed:10000 Mtu:8900} {Name:7e22b55ec9c4989 MacAddress:f6:bd:3e:84:a6:be Speed:10000 Mtu:8900} {Name:7f2b0f7328f1fdf MacAddress:1e:98:1c:a3:fb:11 Speed:10000 Mtu:8900} {Name:8066239534fa93e MacAddress:36:81:fc:cb:ba:ff Speed:10000 Mtu:8900} {Name:80b1488f509c166 MacAddress:52:50:da:89:1c:a3 Speed:10000 Mtu:8900} {Name:8114256db030142 MacAddress:56:1b:41:13:2d:bf Speed:10000 Mtu:8900} {Name:812553092d13019 MacAddress:22:1f:ec:d9:00:a1 Speed:10000 Mtu:8900} {Name:8904f4db7f08be6 MacAddress:da:6c:68:0a:e3:31 Speed:10000 Mtu:8900} {Name:897152fdb47d25f MacAddress:46:5c:d3:ec:8f:37 Speed:10000 Mtu:8900} {Name:89c8e517113e009 MacAddress:b2:60:01:03:12:96 Speed:10000 Mtu:8900} {Name:97c0ac3e4d477ff MacAddress:a2:91:10:23:82:cf Speed:10000 Mtu:8900} {Name:97d00d6aa4d6661 MacAddress:86:35:c5:72:7d:54 Speed:10000 Mtu:8900} {Name:9dab351b1866858 MacAddress:ce:27:07:87:f6:69 Speed:10000 Mtu:8900} {Name:a669846f3b9be13 MacAddress:de:1d:c4:3d:59:80 Speed:10000 Mtu:8900} {Name:aa4c28f3537eda8 MacAddress:2e:fe:3f:e1:6f:a3 Speed:10000 Mtu:8900} {Name:aa871dfa794fea5 MacAddress:9a:92:7b:ac:c9:59 Speed:10000 Mtu:8900} {Name:b7452830de99c04 MacAddress:c2:1d:24:b1:50:ac Speed:10000 Mtu:8900} {Name:ba6a43bb61d242c MacAddress:1e:fa:ee:3c:42:ee Speed:10000 Mtu:8900} {Name:bd5095616b6a849 MacAddress:a2:ef:a3:2a:4f:91 Speed:10000 Mtu:8900} {Name:bdfbd9daf37e036 MacAddress:de:e9:9b:b0:9b:e1 Speed:10000 Mtu:8900} {Name:br-ex MacAddress:fa:16:9e:81:f6:10 Speed:0 Mtu:9000} {Name:br-int MacAddress:d6:69:d9:06:a4:58 Speed:0 Mtu:8900} {Name:c22e02aa77c0fa7 MacAddress:e2:85:0e:7d:f0:79 Speed:10000 Mtu:8900} {Name:c93da22a302a383 MacAddress:1e:ce:73:26:ac:95 Speed:10000 Mtu:8900} {Name:ccbec542fddb9bd MacAddress:4a:ef:f1:f6:11:27 Speed:10000 Mtu:8900} {Name:d4750111f1d0e54 MacAddress:7e:f6:6e:36:78:19 Speed:10000 Mtu:8900} {Name:d996cd9dee5c726 MacAddress:9e:f0:c5:48:9b:cb Speed:10000 Mtu:8900} {Name:e2506cc6607d77e MacAddress:6a:bb:3f:ad:f7:51 Speed:10000 Mtu:8900} {Name:eecf6d3a967ce07 MacAddress:fa:0a:6b:a5:56:79 Speed:10000 Mtu:8900} {Name:eth0 MacAddress:fa:16:9e:81:f6:10 Speed:-1 Mtu:9000} {Name:eth1 MacAddress:fa:16:3e:5e:6c:eb Speed:-1 Mtu:9000} {Name:eth2 MacAddress:fa:16:3e:5e:ad:70 Speed:-1 Mtu:9000} {Name:f12647ed460372c MacAddress:76:b8:00:44:95:c7 Speed:10000 Mtu:8900} {Name:f77a8c44878083f MacAddress:e2:e2:26:88:80:9e Speed:10000 Mtu:8900} {Name:fedfe3c8bf160c8 MacAddress:8a:81:32:e2:c5:80 Speed:10000 Mtu:8900} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:80:00:02 Speed:0 Mtu:8900} {Name:ovs-system MacAddress:6e:04:ec:85:c4:7f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:50514153472 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[12] Caches:[{Id:12 Size:32768 Type:Data Level:1} {Id:12 Size:32768 Type:Instruction Level:1} {Id:12 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:12 Size:16777216 Type:Unified Level:3}] SocketID:12 BookID: DrawerID:} {Id:0 Threads:[13] Caches:[{Id:13 Size:32768 Type:Data Level:1} {Id:13 Size:32768 Type:Instruction Level:1} {Id:13 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:13 Size:16777216 Type:Unified Level:3}] SocketID:13 BookID: DrawerID:} {Id:0 Threads:[14] Caches:[{Id:14 Size:32768 Type:Data Level:1} {Id:14 Size:32768 Type:Instruction Level:1} {Id:14 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:14 Size:16777216 Type:Unified Level:3}] SocketID:14 BookID: DrawerID:} {Id:0 Threads:[15] Caches:[{Id:15 Size:32768 Type:Data Level:1} {Id:15 Size:32768 Type:Instruction Level:1} {Id:15 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:15 Size:16777216 Type:Unified Level:3}] SocketID:15 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 05 10:47:20.118914 master-0 kubenswrapper[24928]: I1205 10:47:20.118147 24928 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 05 10:47:20.118914 master-0 kubenswrapper[24928]: I1205 10:47:20.118229 24928 manager.go:233] Version: {KernelVersion:5.14.0-427.100.1.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202511170715-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 05 10:47:20.118914 master-0 kubenswrapper[24928]: I1205 10:47:20.118681 24928 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 05 10:47:20.119544 master-0 kubenswrapper[24928]: I1205 10:47:20.118898 24928 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 05 10:47:20.119544 master-0 kubenswrapper[24928]: I1205 10:47:20.118947 24928 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"master-0","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"500m","ephemeral-storage":"1Gi","memory":"1Gi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 05 10:47:20.119544 master-0 kubenswrapper[24928]: I1205 10:47:20.119208 24928 topology_manager.go:138] "Creating topology manager with none policy" Dec 05 10:47:20.119544 master-0 kubenswrapper[24928]: I1205 10:47:20.119221 24928 container_manager_linux.go:303] "Creating device plugin manager" Dec 05 10:47:20.119544 master-0 kubenswrapper[24928]: I1205 10:47:20.119232 24928 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 10:47:20.119544 master-0 kubenswrapper[24928]: I1205 10:47:20.119260 24928 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 10:47:20.119544 master-0 kubenswrapper[24928]: I1205 10:47:20.119321 24928 state_mem.go:36] "Initialized new in-memory state store" Dec 05 10:47:20.119544 master-0 kubenswrapper[24928]: I1205 10:47:20.119452 24928 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 05 10:47:20.119544 master-0 kubenswrapper[24928]: I1205 10:47:20.119539 24928 kubelet.go:418] "Attempting to sync node with API server" Dec 05 10:47:20.120006 master-0 kubenswrapper[24928]: I1205 10:47:20.119556 24928 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 05 10:47:20.120006 master-0 kubenswrapper[24928]: I1205 10:47:20.119574 24928 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 05 10:47:20.120006 master-0 kubenswrapper[24928]: I1205 10:47:20.119588 24928 kubelet.go:324] "Adding apiserver pod source" Dec 05 10:47:20.120006 master-0 kubenswrapper[24928]: I1205 10:47:20.119612 24928 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 05 10:47:20.120939 master-0 kubenswrapper[24928]: I1205 10:47:20.120707 24928 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.13-2.rhaos4.18.git15789b8.el9" apiVersion="v1" Dec 05 10:47:20.121566 master-0 kubenswrapper[24928]: I1205 10:47:20.121523 24928 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 05 10:47:20.121909 master-0 kubenswrapper[24928]: I1205 10:47:20.121884 24928 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 05 10:47:20.122087 master-0 kubenswrapper[24928]: I1205 10:47:20.122053 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 05 10:47:20.122391 master-0 kubenswrapper[24928]: I1205 10:47:20.122296 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 05 10:47:20.122487 master-0 kubenswrapper[24928]: I1205 10:47:20.122396 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 05 10:47:20.122487 master-0 kubenswrapper[24928]: I1205 10:47:20.122411 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 05 10:47:20.122487 master-0 kubenswrapper[24928]: I1205 10:47:20.122465 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 05 10:47:20.122487 master-0 kubenswrapper[24928]: I1205 10:47:20.122481 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 05 10:47:20.123907 master-0 kubenswrapper[24928]: I1205 10:47:20.122498 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 05 10:47:20.123907 master-0 kubenswrapper[24928]: I1205 10:47:20.122510 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 05 10:47:20.123907 master-0 kubenswrapper[24928]: I1205 10:47:20.122529 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 05 10:47:20.123907 master-0 kubenswrapper[24928]: I1205 10:47:20.122541 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 05 10:47:20.123907 master-0 kubenswrapper[24928]: I1205 10:47:20.122573 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 05 10:47:20.123907 master-0 kubenswrapper[24928]: I1205 10:47:20.122596 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 05 10:47:20.123907 master-0 kubenswrapper[24928]: I1205 10:47:20.122645 24928 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 05 10:47:20.135393 master-0 kubenswrapper[24928]: I1205 10:47:20.123874 24928 server.go:1280] "Started kubelet" Dec 05 10:47:20.135393 master-0 kubenswrapper[24928]: I1205 10:47:20.125107 24928 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 05 10:47:20.135393 master-0 kubenswrapper[24928]: I1205 10:47:20.125273 24928 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 05 10:47:20.135393 master-0 kubenswrapper[24928]: I1205 10:47:20.125366 24928 server_v1.go:47] "podresources" method="list" useActivePods=true Dec 05 10:47:20.135393 master-0 kubenswrapper[24928]: I1205 10:47:20.125797 24928 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 05 10:47:20.135393 master-0 kubenswrapper[24928]: I1205 10:47:20.127184 24928 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 10:47:20.135393 master-0 kubenswrapper[24928]: I1205 10:47:20.129540 24928 server.go:449] "Adding debug handlers to kubelet server" Dec 05 10:47:20.135393 master-0 kubenswrapper[24928]: I1205 10:47:20.129946 24928 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 10:47:20.134872 master-0 systemd[1]: Started Kubernetes Kubelet. Dec 05 10:47:20.146929 master-0 kubenswrapper[24928]: I1205 10:47:20.146865 24928 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 05 10:47:20.146929 master-0 kubenswrapper[24928]: I1205 10:47:20.146924 24928 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 05 10:47:20.147207 master-0 kubenswrapper[24928]: I1205 10:47:20.146985 24928 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2025-12-06 10:27:25 +0000 UTC, rotation deadline is 2025-12-06 07:29:03.720771989 +0000 UTC Dec 05 10:47:20.147207 master-0 kubenswrapper[24928]: I1205 10:47:20.147073 24928 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 20h41m43.573704871s for next certificate rotation Dec 05 10:47:20.147476 master-0 kubenswrapper[24928]: I1205 10:47:20.147408 24928 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 05 10:47:20.147476 master-0 kubenswrapper[24928]: I1205 10:47:20.147460 24928 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 05 10:47:20.147558 master-0 kubenswrapper[24928]: I1205 10:47:20.147515 24928 desired_state_of_world_populator.go:147] "Desired state populator starts to run" Dec 05 10:47:20.148268 master-0 kubenswrapper[24928]: E1205 10:47:20.148216 24928 kubelet.go:1495] "Image garbage collection failed once. Stats initialization may not have completed yet" err="failed to get imageFs info: unable to find data in memory cache" Dec 05 10:47:20.148583 master-0 kubenswrapper[24928]: I1205 10:47:20.148454 24928 factory.go:55] Registering systemd factory Dec 05 10:47:20.148583 master-0 kubenswrapper[24928]: I1205 10:47:20.148514 24928 factory.go:221] Registration of the systemd container factory successfully Dec 05 10:47:20.149155 master-0 kubenswrapper[24928]: I1205 10:47:20.149119 24928 factory.go:153] Registering CRI-O factory Dec 05 10:47:20.149155 master-0 kubenswrapper[24928]: I1205 10:47:20.149152 24928 factory.go:221] Registration of the crio container factory successfully Dec 05 10:47:20.149324 master-0 kubenswrapper[24928]: I1205 10:47:20.149238 24928 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 05 10:47:20.149324 master-0 kubenswrapper[24928]: I1205 10:47:20.149262 24928 factory.go:103] Registering Raw factory Dec 05 10:47:20.149324 master-0 kubenswrapper[24928]: I1205 10:47:20.149280 24928 manager.go:1196] Started watching for new ooms in manager Dec 05 10:47:20.149830 master-0 kubenswrapper[24928]: I1205 10:47:20.149806 24928 manager.go:319] Starting recovery of all containers Dec 05 10:47:20.151692 master-0 kubenswrapper[24928]: I1205 10:47:20.151648 24928 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 10:47:20.166977 master-0 kubenswrapper[24928]: I1205 10:47:20.166903 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7b29f89-e42d-4e53-ad14-05efdce933f0" volumeName="kubernetes.io/configmap/f7b29f89-e42d-4e53-ad14-05efdce933f0-trusted-ca" seLinuxMountContext="" Dec 05 10:47:20.166977 master-0 kubenswrapper[24928]: I1205 10:47:20.166968 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="11f563d5-89bb-433c-956a-6d5d2492e8f1" volumeName="kubernetes.io/secret/11f563d5-89bb-433c-956a-6d5d2492e8f1-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.166982 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1e69ce9e-4e6f-4015-9ba6-5a7942570190" volumeName="kubernetes.io/secret/1e69ce9e-4e6f-4015-9ba6-5a7942570190-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167000 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="384f1c47-f511-4b58-aa7f-71aef6ef91a9" volumeName="kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167053 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="568e9bed-cbd2-49bd-84af-9e17f95a003f" volumeName="kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-node-bootstrap-token" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167061 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a4fd453c-e667-4bdc-aa9e-3d95ff707200" volumeName="kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-kube-api-access-kmhhw" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167070 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="444f8808-e454-4015-9e20-429e715a08c7" volumeName="kubernetes.io/projected/444f8808-e454-4015-9e20-429e715a08c7-kube-api-access" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167078 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c22d947f-a5b6-4f24-b142-dd201c46293b" volumeName="kubernetes.io/empty-dir/c22d947f-a5b6-4f24-b142-dd201c46293b-available-featuregates" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167091 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0de26264-46c2-4d29-97da-25a1681d6a8e" volumeName="kubernetes.io/projected/0de26264-46c2-4d29-97da-25a1681d6a8e-kube-api-access-cdc7g" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167100 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="21a9bd5a-3ade-48ef-8004-e0492daa85ba" volumeName="kubernetes.io/configmap/21a9bd5a-3ade-48ef-8004-e0492daa85ba-metrics-client-ca" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167109 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="48bd1d86-a6f2-439f-ab04-6a9a442bec42" volumeName="kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cert" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167117 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5a1bdc70-6412-47e0-8330-04d796cc8d55" volumeName="kubernetes.io/projected/5a1bdc70-6412-47e0-8330-04d796cc8d55-kube-api-access-5dtfn" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167125 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-config" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167137 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1fd0a349-0f66-4c85-95e4-8bbc96648c05" volumeName="kubernetes.io/configmap/1fd0a349-0f66-4c85-95e4-8bbc96648c05-service-ca" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167147 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3332b604-6f0d-4243-bce1-b3ab76e58c78" volumeName="kubernetes.io/projected/3332b604-6f0d-4243-bce1-b3ab76e58c78-kube-api-access-6gfgm" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167156 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" volumeName="kubernetes.io/projected/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-kube-api-access-jrdsv" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167166 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="65e06ca6-1f15-43ec-b8dc-a300e1cf83bd" volumeName="kubernetes.io/projected/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-kube-api-access-vtsw9" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167175 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af7078c2-ad4f-415b-ba2f-77f5145c3b3f" volumeName="kubernetes.io/projected/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-kube-api-access-ljdt5" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167184 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01488ac4-313d-48d5-9e86-7b2011b9e91e" volumeName="kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-images" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167541 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3aa9a063-322b-4dc6-a724-05a66026160b" volumeName="kubernetes.io/projected/3aa9a063-322b-4dc6-a724-05a66026160b-kube-api-access-5rlts" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167568 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="445d75af-d072-4fa0-91a7-f3fa579b9ca9" volumeName="kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-default-certificate" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167581 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="88cccb5b-1ad1-4fab-b34c-90252794ee20" volumeName="kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-client-ca" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167590 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af7078c2-ad4f-415b-ba2f-77f5145c3b3f" volumeName="kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-images" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167599 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4c2975ec-e33d-4960-a708-277d41c79b15" volumeName="kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-profile-collector-cert" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167608 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ca45c52e-fb30-4e7c-8c3f-e685c0909916" volumeName="kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-policies" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167618 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3aa9a063-322b-4dc6-a724-05a66026160b" volumeName="kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167685 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="495ba1ea-f844-43ec-8be7-47e738f5428a" volumeName="kubernetes.io/secret/495ba1ea-f844-43ec-8be7-47e738f5428a-ovn-node-metrics-cert" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167699 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="62f174ff-9f3c-4051-a34a-e55a59207171" volumeName="kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-config" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167709 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="65e06ca6-1f15-43ec-b8dc-a300e1cf83bd" volumeName="kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-trusted-ca-bundle" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167720 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d9ff1ae2-8547-42db-9fd0-f4782589ca18" volumeName="kubernetes.io/projected/d9ff1ae2-8547-42db-9fd0-f4782589ca18-kube-api-access-wpgqz" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167729 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="384f1c47-f511-4b58-aa7f-71aef6ef91a9" volumeName="kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167739 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="384f1c47-f511-4b58-aa7f-71aef6ef91a9" volumeName="kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167749 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="88cccb5b-1ad1-4fab-b34c-90252794ee20" volumeName="kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-config" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167757 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ca45c52e-fb30-4e7c-8c3f-e685c0909916" volumeName="kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-serving-ca" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167792 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d5198438-06ae-4e63-a7e3-950ba23bba9c" volumeName="kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-client" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167806 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1a0235af-2cf2-4ad4-b419-764fb56a0107" volumeName="kubernetes.io/secret/1a0235af-2cf2-4ad4-b419-764fb56a0107-metrics-tls" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167816 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b760849c-8d83-47da-8677-68445c143bef" volumeName="kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-ovnkube-config" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167826 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" volumeName="kubernetes.io/projected/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-kube-api-access-95jmp" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167859 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="444f8808-e454-4015-9e20-429e715a08c7" volumeName="kubernetes.io/configmap/444f8808-e454-4015-9e20-429e715a08c7-config" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167868 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fd58232c-a81a-4aee-8b2c-5ffcdded2e23" volumeName="kubernetes.io/configmap/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-config" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167877 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a722cda9-29a0-4b7f-8e1d-9a8950ed765a" volumeName="kubernetes.io/projected/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-kube-api-access-2lwgq" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167885 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b7e7f216-f9af-41df-a0b0-df2a76b9b72a" volumeName="kubernetes.io/secret/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-proxy-tls" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167894 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d5198438-06ae-4e63-a7e3-950ba23bba9c" volumeName="kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-config" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167936 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082" volumeName="kubernetes.io/projected/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-kube-api-access-jv4zs" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167947 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="54f9f9d4-8c31-4636-8010-22ee11b9b323" volumeName="kubernetes.io/empty-dir/54f9f9d4-8c31-4636-8010-22ee11b9b323-volume-directive-shadow" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167962 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="568e9bed-cbd2-49bd-84af-9e17f95a003f" volumeName="kubernetes.io/projected/568e9bed-cbd2-49bd-84af-9e17f95a003f-kube-api-access-flvlq" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167976 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="82464e92-4de2-48f0-8772-a489abb16898" volumeName="kubernetes.io/projected/82464e92-4de2-48f0-8772-a489abb16898-kube-api-access-4dk2b" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167987 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8c649a16-c187-412e-b5da-62a00bee38ab" volumeName="kubernetes.io/projected/8c649a16-c187-412e-b5da-62a00bee38ab-kube-api-access-d8pv2" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.167997 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ee0b4a7f-1b96-4304-bef0-fb575e0e431c" volumeName="kubernetes.io/projected/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-kube-api-access-9ktnb" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168030 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082" volumeName="kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168040 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="65e06ca6-1f15-43ec-b8dc-a300e1cf83bd" volumeName="kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-service-ca-bundle" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168050 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="da9d48c9-6346-4c9f-a690-f7419499c3e6" volumeName="kubernetes.io/projected/da9d48c9-6346-4c9f-a690-f7419499c3e6-kube-api-access-7824l" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168066 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0a4db531-8af7-4085-a6b2-0de51b527ce6" volumeName="kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-utilities" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168076 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="592373ae-a952-4dd3-a9bc-f9c9c19c0802" volumeName="kubernetes.io/empty-dir/592373ae-a952-4dd3-a9bc-f9c9c19c0802-cache" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168088 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6f76d12f-5406-47e2-8337-2f50e35376d6" volumeName="kubernetes.io/secret/6f76d12f-5406-47e2-8337-2f50e35376d6-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168102 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="97192d4d-15d3-4740-82f3-d0d45f9fe7b9" volumeName="kubernetes.io/secret/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-key" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168116 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="24444474-aa9a-4a0f-8b4d-90f0009e0dc7" volumeName="kubernetes.io/secret/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168131 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="283122ba-be1c-4516-bd0f-df41c13c098b" volumeName="kubernetes.io/projected/283122ba-be1c-4516-bd0f-df41c13c098b-kube-api-access-vhjhk" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168144 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4825316a-ea9f-4d3d-838b-fa809a6e49c7" volumeName="kubernetes.io/secret/4825316a-ea9f-4d3d-838b-fa809a6e49c7-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168157 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="489b6771-aa16-469f-9883-9a9233d3c379" volumeName="kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-utilities" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168170 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="58baad85-de54-49e7-a13e-d470d9c50d11" volumeName="kubernetes.io/configmap/58baad85-de54-49e7-a13e-d470d9c50d11-config-volume" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168186 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e" volumeName="kubernetes.io/projected/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-kube-api-access-g88bl" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168198 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ca45c52e-fb30-4e7c-8c3f-e685c0909916" volumeName="kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-trusted-ca-bundle" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168210 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7b29f89-e42d-4e53-ad14-05efdce933f0" volumeName="kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-bound-sa-token" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168223 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="eb290494-a456-4f0e-9afc-f20abab1a1bf" volumeName="kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168237 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1e69ce9e-4e6f-4015-9ba6-5a7942570190" volumeName="kubernetes.io/projected/1e69ce9e-4e6f-4015-9ba6-5a7942570190-kube-api-access-7wfsv" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168249 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="489b6771-aa16-469f-9883-9a9233d3c379" volumeName="kubernetes.io/projected/489b6771-aa16-469f-9883-9a9233d3c379-kube-api-access-9pvrn" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168264 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="baee05cb-62ad-4eda-8a13-e317eae840e5" volumeName="kubernetes.io/projected/baee05cb-62ad-4eda-8a13-e317eae840e5-kube-api-access-564nt" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168276 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" volumeName="kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-proxy-ca-bundles" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168289 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="eb290494-a456-4f0e-9afc-f20abab1a1bf" volumeName="kubernetes.io/configmap/eb290494-a456-4f0e-9afc-f20abab1a1bf-trusted-ca" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168302 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9f011f-36f1-4308-a365-69425c186c7f" volumeName="kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-cni-binary-copy" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168314 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e" volumeName="kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168326 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e27c0798-ec1c-43cd-b81b-f77f2f11ad0f" volumeName="kubernetes.io/projected/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f-kube-api-access-ml74w" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168339 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3aa9a063-322b-4dc6-a724-05a66026160b" volumeName="kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168351 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49051e6e-5a2f-45c8-bad0-374514a91c07" volumeName="kubernetes.io/projected/49051e6e-5a2f-45c8-bad0-374514a91c07-kube-api-access-6vp9b" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168364 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b7e7f216-f9af-41df-a0b0-df2a76b9b72a" volumeName="kubernetes.io/projected/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-kube-api-access-rvw27" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168377 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="baa812e4-b011-41ee-82ab-8f571f0b7e0a" volumeName="kubernetes.io/projected/baa812e4-b011-41ee-82ab-8f571f0b7e0a-kube-api-access-28qvt" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168491 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d5198438-06ae-4e63-a7e3-950ba23bba9c" volumeName="kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-trusted-ca-bundle" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168505 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="384f1c47-f511-4b58-aa7f-71aef6ef91a9" volumeName="kubernetes.io/projected/384f1c47-f511-4b58-aa7f-71aef6ef91a9-kube-api-access-lgtnt" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168520 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="489b6771-aa16-469f-9883-9a9233d3c379" volumeName="kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-catalog-content" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168532 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87909f47-f2d7-46f8-a1c8-27336cdcce5d" volumeName="kubernetes.io/projected/87909f47-f2d7-46f8-a1c8-27336cdcce5d-kube-api-access-rjrgm" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168570 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af7078c2-ad4f-415b-ba2f-77f5145c3b3f" volumeName="kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-auth-proxy-config" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168586 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af7078c2-ad4f-415b-ba2f-77f5145c3b3f" volumeName="kubernetes.io/secret/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-cloud-controller-manager-operator-tls" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168595 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" volumeName="kubernetes.io/secret/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168605 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3332b604-6f0d-4243-bce1-b3ab76e58c78" volumeName="kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168614 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="54f9f9d4-8c31-4636-8010-22ee11b9b323" volumeName="kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-kube-rbac-proxy-config" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168624 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c22d947f-a5b6-4f24-b142-dd201c46293b" volumeName="kubernetes.io/projected/c22d947f-a5b6-4f24-b142-dd201c46293b-kube-api-access-98qt8" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168634 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" volumeName="kubernetes.io/secret/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168644 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d9ff1ae2-8547-42db-9fd0-f4782589ca18" volumeName="kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-tls" seLinuxMountContext="" Dec 05 10:47:20.168582 master-0 kubenswrapper[24928]: I1205 10:47:20.168653 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2c3fab4d-05b8-42ec-8c91-91ff64562649" volumeName="kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-webhook-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168663 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3c4f767d-65f9-434b-8ddd-ceb0b91ab99a" volumeName="kubernetes.io/projected/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a-kube-api-access-pdsh9" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168671 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="97192d4d-15d3-4740-82f3-d0d45f9fe7b9" volumeName="kubernetes.io/projected/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-kube-api-access-nn6sc" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168680 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d5198438-06ae-4e63-a7e3-950ba23bba9c" volumeName="kubernetes.io/projected/d5198438-06ae-4e63-a7e3-950ba23bba9c-kube-api-access-m88gw" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168690 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3aa9a063-322b-4dc6-a724-05a66026160b" volumeName="kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168699 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6728a458-098b-44f9-8c8f-b0d76c5825fa" volumeName="kubernetes.io/configmap/6728a458-098b-44f9-8c8f-b0d76c5825fa-mcc-auth-proxy-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168708 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6f76d12f-5406-47e2-8337-2f50e35376d6" volumeName="kubernetes.io/configmap/6f76d12f-5406-47e2-8337-2f50e35376d6-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168719 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d5198438-06ae-4e63-a7e3-950ba23bba9c" volumeName="kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-image-import-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168728 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7b29f89-e42d-4e53-ad14-05efdce933f0" volumeName="kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-kube-api-access-khfxz" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168738 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0a4db531-8af7-4085-a6b2-0de51b527ce6" volumeName="kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-catalog-content" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168747 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2c3fab4d-05b8-42ec-8c91-91ff64562649" volumeName="kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-apiservice-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168755 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="62f174ff-9f3c-4051-a34a-e55a59207171" volumeName="kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-auth-proxy-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168765 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="97192d4d-15d3-4740-82f3-d0d45f9fe7b9" volumeName="kubernetes.io/configmap/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-cabundle" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168774 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ca45c52e-fb30-4e7c-8c3f-e685c0909916" volumeName="kubernetes.io/projected/ca45c52e-fb30-4e7c-8c3f-e685c0909916-kube-api-access-ltjt6" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168783 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4c2975ec-e33d-4960-a708-277d41c79b15" volumeName="kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-srv-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168798 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01488ac4-313d-48d5-9e86-7b2011b9e91e" volumeName="kubernetes.io/secret/01488ac4-313d-48d5-9e86-7b2011b9e91e-proxy-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168808 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="48bd1d86-a6f2-439f-ab04-6a9a442bec42" volumeName="kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-images" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168817 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="926263c4-ec5b-41cb-9c30-0c88f636035f" volumeName="kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168827 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" volumeName="kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-client-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168837 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5a1bdc70-6412-47e0-8330-04d796cc8d55" volumeName="kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168848 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0aa6e138-4b1f-4047-8255-a2b14d044588" volumeName="kubernetes.io/projected/0aa6e138-4b1f-4047-8255-a2b14d044588-kube-api-access-wnzgs" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168858 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="21a9bd5a-3ade-48ef-8004-e0492daa85ba" volumeName="kubernetes.io/empty-dir/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-textfile" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168869 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="445d75af-d072-4fa0-91a7-f3fa579b9ca9" volumeName="kubernetes.io/configmap/445d75af-d072-4fa0-91a7-f3fa579b9ca9-service-ca-bundle" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168878 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d5198438-06ae-4e63-a7e3-950ba23bba9c" volumeName="kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-serving-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168890 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ee0b4a7f-1b96-4304-bef0-fb575e0e431c" volumeName="kubernetes.io/secret/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cloud-credential-operator-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168900 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168910 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22676fac-b770-4937-9bee-7478bd1babb7" volumeName="kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-kube-api-access-vkqz7" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168920 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2c3fab4d-05b8-42ec-8c91-91ff64562649" volumeName="kubernetes.io/projected/2c3fab4d-05b8-42ec-8c91-91ff64562649-kube-api-access-l8wnp" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168929 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="58b95ae4-7f74-4777-8441-0a0ae28199e9" volumeName="kubernetes.io/projected/58b95ae4-7f74-4777-8441-0a0ae28199e9-kube-api-access-7fbtj" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168938 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="62f174ff-9f3c-4051-a34a-e55a59207171" volumeName="kubernetes.io/secret/62f174ff-9f3c-4051-a34a-e55a59207171-machine-approver-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168948 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="720a7467-ce93-4d48-82ec-9ad0922d99c2" volumeName="kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168957 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="eb290494-a456-4f0e-9afc-f20abab1a1bf" volumeName="kubernetes.io/projected/eb290494-a456-4f0e-9afc-f20abab1a1bf-kube-api-access-qn7rj" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168969 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1f132702-304b-46ff-b428-6df1deeffec3" volumeName="kubernetes.io/projected/1f132702-304b-46ff-b428-6df1deeffec3-kube-api-access-vkzkh" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168980 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" volumeName="kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.168994 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3aa9a063-322b-4dc6-a724-05a66026160b" volumeName="kubernetes.io/empty-dir/3aa9a063-322b-4dc6-a724-05a66026160b-audit-log" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169006 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082" volumeName="kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-profile-collector-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169022 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" volumeName="kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-service-ca-bundle" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169037 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169051 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1a0235af-2cf2-4ad4-b419-764fb56a0107" volumeName="kubernetes.io/projected/1a0235af-2cf2-4ad4-b419-764fb56a0107-kube-api-access-cwqkb" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169140 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="21a9bd5a-3ade-48ef-8004-e0492daa85ba" volumeName="kubernetes.io/projected/21a9bd5a-3ade-48ef-8004-e0492daa85ba-kube-api-access-r8bpw" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169152 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="33616deb-ea10-4a38-8681-ab023b526b11" volumeName="kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-etc-tuned" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169161 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c" volumeName="kubernetes.io/projected/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-kube-api-access-ssssf" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169172 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d9ff1ae2-8547-42db-9fd0-f4782589ca18" volumeName="kubernetes.io/configmap/d9ff1ae2-8547-42db-9fd0-f4782589ca18-metrics-client-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169183 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0aa6e138-4b1f-4047-8255-a2b14d044588" volumeName="kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-utilities" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169193 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" volumeName="kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-binary-copy" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169202 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-client" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169211 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01488ac4-313d-48d5-9e86-7b2011b9e91e" volumeName="kubernetes.io/projected/01488ac4-313d-48d5-9e86-7b2011b9e91e-kube-api-access-2xjx8" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169220 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b7508cd3-7421-4ab9-be1f-318db2853596" volumeName="kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169229 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d9ff1ae2-8547-42db-9fd0-f4782589ca18" volumeName="kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-kube-rbac-proxy-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169238 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0aa6e138-4b1f-4047-8255-a2b14d044588" volumeName="kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-catalog-content" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169247 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4c2975ec-e33d-4960-a708-277d41c79b15" volumeName="kubernetes.io/projected/4c2975ec-e33d-4960-a708-277d41c79b15-kube-api-access-d6z8n" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169256 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6728a458-098b-44f9-8c8f-b0d76c5825fa" volumeName="kubernetes.io/secret/6728a458-098b-44f9-8c8f-b0d76c5825fa-proxy-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169275 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22676fac-b770-4937-9bee-7478bd1babb7" volumeName="kubernetes.io/configmap/22676fac-b770-4937-9bee-7478bd1babb7-trusted-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169285 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22676fac-b770-4937-9bee-7478bd1babb7" volumeName="kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169293 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e" volumeName="kubernetes.io/secret/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-webhook-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169303 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4825316a-ea9f-4d3d-838b-fa809a6e49c7" volumeName="kubernetes.io/projected/4825316a-ea9f-4d3d-838b-fa809a6e49c7-kube-api-access" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169312 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="54f9f9d4-8c31-4636-8010-22ee11b9b323" volumeName="kubernetes.io/projected/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-api-access-sv2tg" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169320 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="82464e92-4de2-48f0-8772-a489abb16898" volumeName="kubernetes.io/secret/82464e92-4de2-48f0-8772-a489abb16898-machine-api-operator-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169329 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a4fd453c-e667-4bdc-aa9e-3d95ff707200" volumeName="kubernetes.io/empty-dir/a4fd453c-e667-4bdc-aa9e-3d95ff707200-cache" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169338 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d5198438-06ae-4e63-a7e3-950ba23bba9c" volumeName="kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169347 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1e69ce9e-4e6f-4015-9ba6-5a7942570190" volumeName="kubernetes.io/configmap/1e69ce9e-4e6f-4015-9ba6-5a7942570190-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169356 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" volumeName="kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-sysctl-allowlist" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169365 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="445d75af-d072-4fa0-91a7-f3fa579b9ca9" volumeName="kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-metrics-certs" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169375 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ee0b4a7f-1b96-4304-bef0-fb575e0e431c" volumeName="kubernetes.io/configmap/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cco-trusted-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169385 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="384f1c47-f511-4b58-aa7f-71aef6ef91a9" volumeName="kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-metrics-client-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169395 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8c649a16-c187-412e-b5da-62a00bee38ab" volumeName="kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169404 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d1c3b7dd-f25e-4983-8a94-084f863fd5b9" volumeName="kubernetes.io/projected/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-kube-api-access-fkmvj" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169415 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e" volumeName="kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-ovnkube-identity-cm" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169448 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="495ba1ea-f844-43ec-8be7-47e738f5428a" volumeName="kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169459 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d1c3b7dd-f25e-4983-8a94-084f863fd5b9" volumeName="kubernetes.io/secret/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169468 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" volumeName="kubernetes.io/projected/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-kube-api-access-2fbss" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169479 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" volumeName="kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-trusted-ca-bundle" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169488 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fd58232c-a81a-4aee-8b2c-5ffcdded2e23" volumeName="kubernetes.io/projected/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-kube-api-access-h2vd4" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169498 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22676fac-b770-4937-9bee-7478bd1babb7" volumeName="kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-bound-sa-token" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169834 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2c3fab4d-05b8-42ec-8c91-91ff64562649" volumeName="kubernetes.io/empty-dir/2c3fab4d-05b8-42ec-8c91-91ff64562649-tmpfs" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169856 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="48bd1d86-a6f2-439f-ab04-6a9a442bec42" volumeName="kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cluster-baremetal-operator-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169867 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="58baad85-de54-49e7-a13e-d470d9c50d11" volumeName="kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169877 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b760849c-8d83-47da-8677-68445c143bef" volumeName="kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-env-overrides" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169889 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01488ac4-313d-48d5-9e86-7b2011b9e91e" volumeName="kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-auth-proxy-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169898 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="592373ae-a952-4dd3-a9bc-f9c9c19c0802" volumeName="kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-kube-api-access-rd87p" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169907 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a4fd453c-e667-4bdc-aa9e-3d95ff707200" volumeName="kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-ca-certs" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169917 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="82464e92-4de2-48f0-8772-a489abb16898" volumeName="kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169926 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a4fd453c-e667-4bdc-aa9e-3d95ff707200" volumeName="kubernetes.io/secret/a4fd453c-e667-4bdc-aa9e-3d95ff707200-catalogserver-certs" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169935 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c22d947f-a5b6-4f24-b142-dd201c46293b" volumeName="kubernetes.io/secret/c22d947f-a5b6-4f24-b142-dd201c46293b-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169946 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e" volumeName="kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-env-overrides" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169955 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fd58232c-a81a-4aee-8b2c-5ffcdded2e23" volumeName="kubernetes.io/secret/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169963 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="283122ba-be1c-4516-bd0f-df41c13c098b" volumeName="kubernetes.io/configmap/283122ba-be1c-4516-bd0f-df41c13c098b-iptables-alerter-script" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169974 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3aa9a063-322b-4dc6-a724-05a66026160b" volumeName="kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.169984 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b760849c-8d83-47da-8677-68445c143bef" volumeName="kubernetes.io/secret/b760849c-8d83-47da-8677-68445c143bef-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170013 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9f011f-36f1-4308-a365-69425c186c7f" volumeName="kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-multus-daemon-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170027 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="baa812e4-b011-41ee-82ab-8f571f0b7e0a" volumeName="kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-utilities" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170036 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="444f8808-e454-4015-9e20-429e715a08c7" volumeName="kubernetes.io/secret/444f8808-e454-4015-9e20-429e715a08c7-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170046 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9f011f-36f1-4308-a365-69425c186c7f" volumeName="kubernetes.io/projected/6a9f011f-36f1-4308-a365-69425c186c7f-kube-api-access-mvnxf" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170056 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b760849c-8d83-47da-8677-68445c143bef" volumeName="kubernetes.io/projected/b760849c-8d83-47da-8677-68445c143bef-kube-api-access-jfz87" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170066 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ca45c52e-fb30-4e7c-8c3f-e685c0909916" volumeName="kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-encryption-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170079 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="495ba1ea-f844-43ec-8be7-47e738f5428a" volumeName="kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-env-overrides" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170090 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d5198438-06ae-4e63-a7e3-950ba23bba9c" volumeName="kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170102 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e" volumeName="kubernetes.io/projected/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-kube-api-access-wg9zq" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170115 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-service-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170129 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1f132702-304b-46ff-b428-6df1deeffec3" volumeName="kubernetes.io/secret/1f132702-304b-46ff-b428-6df1deeffec3-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170146 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="384f1c47-f511-4b58-aa7f-71aef6ef91a9" volumeName="kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170160 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="445d75af-d072-4fa0-91a7-f3fa579b9ca9" volumeName="kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-stats-auth" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170171 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a722cda9-29a0-4b7f-8e1d-9a8950ed765a" volumeName="kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170182 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e" volumeName="kubernetes.io/configmap/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-metrics-client-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170191 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="65e06ca6-1f15-43ec-b8dc-a300e1cf83bd" volumeName="kubernetes.io/empty-dir/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-snapshots" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170199 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7b29f89-e42d-4e53-ad14-05efdce933f0" volumeName="kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170208 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3aa9a063-322b-4dc6-a724-05a66026160b" volumeName="kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170219 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4825316a-ea9f-4d3d-838b-fa809a6e49c7" volumeName="kubernetes.io/configmap/4825316a-ea9f-4d3d-838b-fa809a6e49c7-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170228 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="48bd1d86-a6f2-439f-ab04-6a9a442bec42" volumeName="kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170237 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="58baad85-de54-49e7-a13e-d470d9c50d11" volumeName="kubernetes.io/projected/58baad85-de54-49e7-a13e-d470d9c50d11-kube-api-access-jx7jr" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170245 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="65e06ca6-1f15-43ec-b8dc-a300e1cf83bd" volumeName="kubernetes.io/secret/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170255 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" volumeName="kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170264 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0a4db531-8af7-4085-a6b2-0de51b527ce6" volumeName="kubernetes.io/projected/0a4db531-8af7-4085-a6b2-0de51b527ce6-kube-api-access-g4jgc" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170273 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0de26264-46c2-4d29-97da-25a1681d6a8e" volumeName="kubernetes.io/secret/0de26264-46c2-4d29-97da-25a1681d6a8e-samples-operator-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170282 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1f132702-304b-46ff-b428-6df1deeffec3" volumeName="kubernetes.io/configmap/1f132702-304b-46ff-b428-6df1deeffec3-auth-proxy-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170292 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="33616deb-ea10-4a38-8681-ab023b526b11" volumeName="kubernetes.io/projected/33616deb-ea10-4a38-8681-ab023b526b11-kube-api-access-rsrkp" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170303 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e" volumeName="kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-kube-rbac-proxy-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170313 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="54f9f9d4-8c31-4636-8010-22ee11b9b323" volumeName="kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-metrics-client-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170322 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="54f9f9d4-8c31-4636-8010-22ee11b9b323" volumeName="kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170480 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="592373ae-a952-4dd3-a9bc-f9c9c19c0802" volumeName="kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-ca-certs" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170515 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="eb290494-a456-4f0e-9afc-f20abab1a1bf" volumeName="kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170527 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="11f563d5-89bb-433c-956a-6d5d2492e8f1" volumeName="kubernetes.io/projected/11f563d5-89bb-433c-956a-6d5d2492e8f1-kube-api-access" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170538 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="82464e92-4de2-48f0-8772-a489abb16898" volumeName="kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-images" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170552 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a722cda9-29a0-4b7f-8e1d-9a8950ed765a" volumeName="kubernetes.io/configmap/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-telemetry-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170562 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0babac16-164a-405a-a86c-30524118f046" volumeName="kubernetes.io/secret/0babac16-164a-405a-a86c-30524118f046-tls-certificates" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170571 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="568e9bed-cbd2-49bd-84af-9e17f95a003f" volumeName="kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-certs" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170618 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6728a458-098b-44f9-8c8f-b0d76c5825fa" volumeName="kubernetes.io/projected/6728a458-098b-44f9-8c8f-b0d76c5825fa-kube-api-access-784mb" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170634 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ca45c52e-fb30-4e7c-8c3f-e685c0909916" volumeName="kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-client" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170648 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1fd0a349-0f66-4c85-95e4-8bbc96648c05" volumeName="kubernetes.io/secret/1fd0a349-0f66-4c85-95e4-8bbc96648c05-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170659 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="384f1c47-f511-4b58-aa7f-71aef6ef91a9" volumeName="kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170669 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="54f9f9d4-8c31-4636-8010-22ee11b9b323" volumeName="kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-custom-resource-state-configmap" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170679 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="384f1c47-f511-4b58-aa7f-71aef6ef91a9" volumeName="kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170692 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49051e6e-5a2f-45c8-bad0-374514a91c07" volumeName="kubernetes.io/empty-dir/49051e6e-5a2f-45c8-bad0-374514a91c07-operand-assets" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170702 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="58b95ae4-7f74-4777-8441-0a0ae28199e9" volumeName="kubernetes.io/secret/58b95ae4-7f74-4777-8441-0a0ae28199e9-cluster-storage-operator-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170713 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="88cccb5b-1ad1-4fab-b34c-90252794ee20" volumeName="kubernetes.io/projected/88cccb5b-1ad1-4fab-b34c-90252794ee20-kube-api-access-f4lbg" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170723 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="baa812e4-b011-41ee-82ab-8f571f0b7e0a" volumeName="kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-catalog-content" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170755 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="21a9bd5a-3ade-48ef-8004-e0492daa85ba" volumeName="kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-tls" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170764 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="445d75af-d072-4fa0-91a7-f3fa579b9ca9" volumeName="kubernetes.io/projected/445d75af-d072-4fa0-91a7-f3fa579b9ca9-kube-api-access-hskm2" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170773 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="495ba1ea-f844-43ec-8be7-47e738f5428a" volumeName="kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-script-lib" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170802 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d5198438-06ae-4e63-a7e3-950ba23bba9c" volumeName="kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-encryption-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170814 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="21a9bd5a-3ade-48ef-8004-e0492daa85ba" volumeName="kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-kube-rbac-proxy-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170822 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49051e6e-5a2f-45c8-bad0-374514a91c07" volumeName="kubernetes.io/secret/49051e6e-5a2f-45c8-bad0-374514a91c07-cluster-olm-operator-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170831 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="88cccb5b-1ad1-4fab-b34c-90252794ee20" volumeName="kubernetes.io/secret/88cccb5b-1ad1-4fab-b34c-90252794ee20-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170841 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="48bd1d86-a6f2-439f-ab04-6a9a442bec42" volumeName="kubernetes.io/projected/48bd1d86-a6f2-439f-ab04-6a9a442bec42-kube-api-access-f69hj" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170850 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="926263c4-ec5b-41cb-9c30-0c88f636035f" volumeName="kubernetes.io/projected/926263c4-ec5b-41cb-9c30-0c88f636035f-kube-api-access-fncz7" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170858 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ca45c52e-fb30-4e7c-8c3f-e685c0909916" volumeName="kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-serving-cert" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170867 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" volumeName="kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170900 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1fd0a349-0f66-4c85-95e4-8bbc96648c05" volumeName="kubernetes.io/projected/1fd0a349-0f66-4c85-95e4-8bbc96648c05-kube-api-access" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170916 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="495ba1ea-f844-43ec-8be7-47e738f5428a" volumeName="kubernetes.io/projected/495ba1ea-f844-43ec-8be7-47e738f5428a-kube-api-access-tjgc4" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170924 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="62f174ff-9f3c-4051-a34a-e55a59207171" volumeName="kubernetes.io/projected/62f174ff-9f3c-4051-a34a-e55a59207171-kube-api-access-r5p8s" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170933 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b7e7f216-f9af-41df-a0b0-df2a76b9b72a" volumeName="kubernetes.io/configmap/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-mcd-auth-proxy-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170941 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d1c3b7dd-f25e-4983-8a94-084f863fd5b9" volumeName="kubernetes.io/configmap/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170950 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="11f563d5-89bb-433c-956a-6d5d2492e8f1" volumeName="kubernetes.io/configmap/11f563d5-89bb-433c-956a-6d5d2492e8f1-config" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170959 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" volumeName="kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-whereabouts-configmap" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170970 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6f76d12f-5406-47e2-8337-2f50e35376d6" volumeName="kubernetes.io/projected/6f76d12f-5406-47e2-8337-2f50e35376d6-kube-api-access-fzwfq" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170979 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="24444474-aa9a-4a0f-8b4d-90f0009e0dc7" volumeName="kubernetes.io/projected/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-kube-api-access-lrwt6" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170989 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="33616deb-ea10-4a38-8681-ab023b526b11" volumeName="kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-tmp" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.170998 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="926263c4-ec5b-41cb-9c30-0c88f636035f" volumeName="kubernetes.io/configmap/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-trusted-ca" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.171007 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b7508cd3-7421-4ab9-be1f-318db2853596" volumeName="kubernetes.io/projected/b7508cd3-7421-4ab9-be1f-318db2853596-kube-api-access-kgvv6" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.171016 24928 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7a08359-0379-4364-8b0c-ddb58ff605f4" volumeName="kubernetes.io/projected/f7a08359-0379-4364-8b0c-ddb58ff605f4-kube-api-access-twh85" seLinuxMountContext="" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.171024 24928 reconstruct.go:97] "Volume reconstruction finished" Dec 05 10:47:20.172183 master-0 kubenswrapper[24928]: I1205 10:47:20.171031 24928 reconciler.go:26] "Reconciler: start to sync state" Dec 05 10:47:20.179310 master-0 kubenswrapper[24928]: I1205 10:47:20.175499 24928 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 05 10:47:20.200483 master-0 kubenswrapper[24928]: I1205 10:47:20.200368 24928 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 05 10:47:20.202955 master-0 kubenswrapper[24928]: I1205 10:47:20.202913 24928 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 05 10:47:20.203040 master-0 kubenswrapper[24928]: I1205 10:47:20.202977 24928 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 05 10:47:20.203040 master-0 kubenswrapper[24928]: I1205 10:47:20.203012 24928 kubelet.go:2335] "Starting kubelet main sync loop" Dec 05 10:47:20.203101 master-0 kubenswrapper[24928]: E1205 10:47:20.203080 24928 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 05 10:47:20.205183 master-0 kubenswrapper[24928]: I1205 10:47:20.205133 24928 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 10:47:20.214110 master-0 kubenswrapper[24928]: I1205 10:47:20.213915 24928 generic.go:334] "Generic (PLEG): container finished" podID="4825316a-ea9f-4d3d-838b-fa809a6e49c7" containerID="f2785fc5d00b92168028cf4e3273a28c6085e8db64c579bdce5eca572512c7d9" exitCode=0 Dec 05 10:47:20.216780 master-0 kubenswrapper[24928]: I1205 10:47:20.216727 24928 generic.go:334] "Generic (PLEG): container finished" podID="21a9bd5a-3ade-48ef-8004-e0492daa85ba" containerID="8ede0c38534b9efcdf0a87cdc9cd4d9eec594a788625f22f913ffac53544f49c" exitCode=0 Dec 05 10:47:20.222351 master-0 kubenswrapper[24928]: I1205 10:47:20.222310 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager-operator_openshift-controller-manager-operator-6c8676f99d-cwvk5_1e69ce9e-4e6f-4015-9ba6-5a7942570190/openshift-controller-manager-operator/1.log" Dec 05 10:47:20.222491 master-0 kubenswrapper[24928]: I1205 10:47:20.222351 24928 generic.go:334] "Generic (PLEG): container finished" podID="1e69ce9e-4e6f-4015-9ba6-5a7942570190" containerID="ded2f19eb6b7e89481ac839fb3ba5dd47ee60cff52debe8dba17401c93d0ac1a" exitCode=255 Dec 05 10:47:20.225583 master-0 kubenswrapper[24928]: I1205 10:47:20.225527 24928 generic.go:334] "Generic (PLEG): container finished" podID="6f76d12f-5406-47e2-8337-2f50e35376d6" containerID="cc8433167039cff15f2a8553e5790127434ae4c5286a5db0777d417011f80f8c" exitCode=0 Dec 05 10:47:20.252053 master-0 kubenswrapper[24928]: I1205 10:47:20.251475 24928 generic.go:334] "Generic (PLEG): container finished" podID="ca45c52e-fb30-4e7c-8c3f-e685c0909916" containerID="0bd0566a23e122fff36d9dfda48e7737f21d3d5b04a9764846004be6154bdb8d" exitCode=0 Dec 05 10:47:20.255154 master-0 kubenswrapper[24928]: I1205 10:47:20.255117 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-catalogd_catalogd-controller-manager-7cc89f4c4c-lth87_a4fd453c-e667-4bdc-aa9e-3d95ff707200/manager/0.log" Dec 05 10:47:20.255435 master-0 kubenswrapper[24928]: I1205 10:47:20.255383 24928 generic.go:334] "Generic (PLEG): container finished" podID="a4fd453c-e667-4bdc-aa9e-3d95ff707200" containerID="2e9155fd74fd1d13dcef2794db107599bb6fa96cc533e3303f4bbbfdbd8c0208" exitCode=1 Dec 05 10:47:20.256375 master-0 kubenswrapper[24928]: I1205 10:47:20.256337 24928 generic.go:334] "Generic (PLEG): container finished" podID="b3c5922c-65b0-4c75-992c-f9c88f81762d" containerID="f02bbe84a0ab6b5c3f31276301465d91b619a21b309cbdf880fc6c1df1dcf605" exitCode=0 Dec 05 10:47:20.258723 master-0 kubenswrapper[24928]: I1205 10:47:20.258693 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-node-identity_network-node-identity-ql7j7_d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/approver/0.log" Dec 05 10:47:20.259177 master-0 kubenswrapper[24928]: I1205 10:47:20.259146 24928 generic.go:334] "Generic (PLEG): container finished" podID="d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e" containerID="5674d03e599731519f29f243f076ae28159025706182191e04e4bbfef0819512" exitCode=1 Dec 05 10:47:20.265029 master-0 kubenswrapper[24928]: I1205 10:47:20.264960 24928 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="12675055319a88464b4e8137b9c8f20eb9ba5b578bde42cad92050f544a0c6be" exitCode=0 Dec 05 10:47:20.265029 master-0 kubenswrapper[24928]: I1205 10:47:20.265023 24928 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="ee9e5b75d34cda6a745ee4f00c4f147db05bd9943033e03d413d4fa85d63bb34" exitCode=0 Dec 05 10:47:20.265029 master-0 kubenswrapper[24928]: I1205 10:47:20.265034 24928 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="3acdc4bc20dace9999e3e38bd13059d0977882b304881b0e3ccd402e31f993c2" exitCode=0 Dec 05 10:47:20.265198 master-0 kubenswrapper[24928]: I1205 10:47:20.265044 24928 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="f4f209477f798940668277c240fd1326e8fd7d7c2b8eedd2830e59be706883e8" exitCode=0 Dec 05 10:47:20.265198 master-0 kubenswrapper[24928]: I1205 10:47:20.265053 24928 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="215c293fe494f6f41ad17e025bbeffa977136f630518896e92b337f0d84c340a" exitCode=0 Dec 05 10:47:20.265198 master-0 kubenswrapper[24928]: I1205 10:47:20.265061 24928 generic.go:334] "Generic (PLEG): container finished" podID="38ad6e6a-a2b9-44e9-ac0b-e413c65efad8" containerID="10b806d181204c81aa236ce7e2f93aa28932b9c9f02c11fefdb2f0d1d7bdeac9" exitCode=0 Dec 05 10:47:20.268200 master-0 kubenswrapper[24928]: I1205 10:47:20.268158 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-operator_network-operator-79767b7ff9-t8j2j_1a0235af-2cf2-4ad4-b419-764fb56a0107/network-operator/0.log" Dec 05 10:47:20.268295 master-0 kubenswrapper[24928]: I1205 10:47:20.268215 24928 generic.go:334] "Generic (PLEG): container finished" podID="1a0235af-2cf2-4ad4-b419-764fb56a0107" containerID="f1e6d80e2ae7d337901560dde2050e55f54f6b182ab45a5e1cce9ebc5f4043a9" exitCode=255 Dec 05 10:47:20.270651 master-0 kubenswrapper[24928]: I1205 10:47:20.270567 24928 generic.go:334] "Generic (PLEG): container finished" podID="e149cbe7-1ec3-4a06-af61-3a8906b8e9ef" containerID="86721e833ddfe5b80f087deb054f43ae1d6033b98f09d3f999eb1efccfb11215" exitCode=0 Dec 05 10:47:20.272400 master-0 kubenswrapper[24928]: I1205 10:47:20.272356 24928 generic.go:334] "Generic (PLEG): container finished" podID="7db4c690-9d13-4c82-9565-f974d554916b" containerID="2abcdc2c255198aabeec8ea169e4ae237c27c37467559f6c930cf9b54982525d" exitCode=0 Dec 05 10:47:20.279712 master-0 kubenswrapper[24928]: I1205 10:47:20.279662 24928 generic.go:334] "Generic (PLEG): container finished" podID="489b6771-aa16-469f-9883-9a9233d3c379" containerID="0a06bffef31dc5f888e335796830f40b7d25388d95911101d91480b1460e1c2e" exitCode=0 Dec 05 10:47:20.279712 master-0 kubenswrapper[24928]: I1205 10:47:20.279704 24928 generic.go:334] "Generic (PLEG): container finished" podID="489b6771-aa16-469f-9883-9a9233d3c379" containerID="8338111de9d79599f0bf5e87a8b5d537443ba25abd595ae9a7a16a59ac9bb008" exitCode=0 Dec 05 10:47:20.284778 master-0 kubenswrapper[24928]: I1205 10:47:20.284734 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-1-master-0_5d8c854f-31f8-42d3-b28c-82c7d1b14ee9/installer/0.log" Dec 05 10:47:20.284855 master-0 kubenswrapper[24928]: I1205 10:47:20.284786 24928 generic.go:334] "Generic (PLEG): container finished" podID="5d8c854f-31f8-42d3-b28c-82c7d1b14ee9" containerID="0d4c12890814d7bd72e2d33d03a6131a3d52c0e2122d3701215dc96f26a231b2" exitCode=1 Dec 05 10:47:20.288767 master-0 kubenswrapper[24928]: I1205 10:47:20.288716 24928 generic.go:334] "Generic (PLEG): container finished" podID="c22d947f-a5b6-4f24-b142-dd201c46293b" containerID="3e549526f5057891751a60c691572ec799806d7061304478b0c1e902a36e6208" exitCode=0 Dec 05 10:47:20.288841 master-0 kubenswrapper[24928]: I1205 10:47:20.288799 24928 generic.go:334] "Generic (PLEG): container finished" podID="c22d947f-a5b6-4f24-b142-dd201c46293b" containerID="d0d4427b1c533edfc1d254b4b4202c620d108a1b343169a0fecb4359d771c631" exitCode=0 Dec 05 10:47:20.301223 master-0 kubenswrapper[24928]: I1205 10:47:20.301154 24928 generic.go:334] "Generic (PLEG): container finished" podID="c24e01603234fe8003f8aae8171b0065" containerID="41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2" exitCode=0 Dec 05 10:47:20.301223 master-0 kubenswrapper[24928]: I1205 10:47:20.301199 24928 generic.go:334] "Generic (PLEG): container finished" podID="c24e01603234fe8003f8aae8171b0065" containerID="d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc" exitCode=0 Dec 05 10:47:20.301223 master-0 kubenswrapper[24928]: I1205 10:47:20.301208 24928 generic.go:334] "Generic (PLEG): container finished" podID="c24e01603234fe8003f8aae8171b0065" containerID="f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2" exitCode=0 Dec 05 10:47:20.303168 master-0 kubenswrapper[24928]: E1205 10:47:20.303137 24928 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 05 10:47:20.307078 master-0 kubenswrapper[24928]: I1205 10:47:20.307041 24928 generic.go:334] "Generic (PLEG): container finished" podID="495ba1ea-f844-43ec-8be7-47e738f5428a" containerID="f02b670dd80cfe6eb6e6f824971df1d4d23971943a981cfcb563de519950bb3e" exitCode=0 Dec 05 10:47:20.310501 master-0 kubenswrapper[24928]: I1205 10:47:20.310456 24928 generic.go:334] "Generic (PLEG): container finished" podID="11f563d5-89bb-433c-956a-6d5d2492e8f1" containerID="475c118c4e82c9f8f86c3034eb04eddc9ae813cf612425cd157e214036c29617" exitCode=0 Dec 05 10:47:20.313467 master-0 kubenswrapper[24928]: I1205 10:47:20.313405 24928 generic.go:334] "Generic (PLEG): container finished" podID="bdd20c68f1ccfd1240fb273af0638d01" containerID="baa0fb92990e2611d666dd15d06398cf84d61306c4b64819cbec0b7a2dd1f1f8" exitCode=0 Dec 05 10:47:20.321398 master-0 kubenswrapper[24928]: I1205 10:47:20.321360 24928 generic.go:334] "Generic (PLEG): container finished" podID="d1c3b7dd-f25e-4983-8a94-084f863fd5b9" containerID="e562ce0da1077d81598e22070b676b5642a7e1f7ad3053d9875573b9c4c32d50" exitCode=0 Dec 05 10:47:20.331383 master-0 kubenswrapper[24928]: I1205 10:47:20.331325 24928 generic.go:334] "Generic (PLEG): container finished" podID="1b60e9f1-b49e-4804-b56d-857c7bb2a18b" containerID="96138a92afe8bfdf53592444354e7fbbf1f63d3a7a1a9363fd5b53fd0a427c72" exitCode=0 Dec 05 10:47:20.336617 master-0 kubenswrapper[24928]: I1205 10:47:20.336565 24928 generic.go:334] "Generic (PLEG): container finished" podID="d5198438-06ae-4e63-a7e3-950ba23bba9c" containerID="4aeba45988671d42c201d0e79c4b96e55ead796c0fe34250d02d4c4bd0c61ad7" exitCode=0 Dec 05 10:47:20.339346 master-0 kubenswrapper[24928]: I1205 10:47:20.339286 24928 generic.go:334] "Generic (PLEG): container finished" podID="49051e6e-5a2f-45c8-bad0-374514a91c07" containerID="8a2b637ea3b720bb56a1f54c7baf3eb708da58746d9f80b01f4fc142a426b6b4" exitCode=0 Dec 05 10:47:20.339346 master-0 kubenswrapper[24928]: I1205 10:47:20.339333 24928 generic.go:334] "Generic (PLEG): container finished" podID="49051e6e-5a2f-45c8-bad0-374514a91c07" containerID="be912c90bc4a342d66fa17918d807831fa70aa4a080f3751a563cdc9efce0a37" exitCode=0 Dec 05 10:47:20.343695 master-0 kubenswrapper[24928]: I1205 10:47:20.343649 24928 generic.go:334] "Generic (PLEG): container finished" podID="f7a08359-0379-4364-8b0c-ddb58ff605f4" containerID="7f719d929cd8f822ace129247877e00bae5a9457d73d9da2566b48784c2ee8fc" exitCode=0 Dec 05 10:47:20.348223 master-0 kubenswrapper[24928]: I1205 10:47:20.348188 24928 generic.go:334] "Generic (PLEG): container finished" podID="0a4db531-8af7-4085-a6b2-0de51b527ce6" containerID="ef5e388b6cabad318ee03ac9a678f15288ffd67d2a57ef3139d9a610725fcfd3" exitCode=0 Dec 05 10:47:20.348223 master-0 kubenswrapper[24928]: I1205 10:47:20.348224 24928 generic.go:334] "Generic (PLEG): container finished" podID="0a4db531-8af7-4085-a6b2-0de51b527ce6" containerID="b5d5f4343a07894afd8f108c3807a5c794b99b06442a88e2827b3d948daa8970" exitCode=0 Dec 05 10:47:20.350889 master-0 kubenswrapper[24928]: I1205 10:47:20.350844 24928 generic.go:334] "Generic (PLEG): container finished" podID="444f8808-e454-4015-9e20-429e715a08c7" containerID="a4e599b9d9dac0eea61bdc89656e6a7ee18f83aa3fd9ba9340f57da02bd4a39b" exitCode=0 Dec 05 10:47:20.353049 master-0 kubenswrapper[24928]: I1205 10:47:20.353016 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler_installer-3-master-0_6325f9a2-86d5-4fac-9e71-331b2628f965/installer/0.log" Dec 05 10:47:20.353149 master-0 kubenswrapper[24928]: I1205 10:47:20.353073 24928 generic.go:334] "Generic (PLEG): container finished" podID="6325f9a2-86d5-4fac-9e71-331b2628f965" containerID="df5bdcc899d9a7c7268a300825923259315bb7a02282ed76b8b7644641b06995" exitCode=1 Dec 05 10:47:20.359008 master-0 kubenswrapper[24928]: I1205 10:47:20.358899 24928 generic.go:334] "Generic (PLEG): container finished" podID="a51fa80258077cf0f6fddf343b4ba54f" containerID="15c71c8a48ba2bacfb2bd2a3e4d67fbc729bbcb6ae32fc023ae61c538c7d05e9" exitCode=0 Dec 05 10:47:20.359008 master-0 kubenswrapper[24928]: I1205 10:47:20.359003 24928 generic.go:334] "Generic (PLEG): container finished" podID="a51fa80258077cf0f6fddf343b4ba54f" containerID="0c2a956260655c961495c5a173ae1d7bbf307d5b8972db5af276ccbded5567cc" exitCode=2 Dec 05 10:47:20.359008 master-0 kubenswrapper[24928]: I1205 10:47:20.359012 24928 generic.go:334] "Generic (PLEG): container finished" podID="a51fa80258077cf0f6fddf343b4ba54f" containerID="f19decd176069e2f8d9129c470cba4222beb8bbea5c63430032cfc884efc92be" exitCode=0 Dec 05 10:47:20.359217 master-0 kubenswrapper[24928]: I1205 10:47:20.359022 24928 generic.go:334] "Generic (PLEG): container finished" podID="a51fa80258077cf0f6fddf343b4ba54f" containerID="b51c377684aafed12957958fdc2a947c34fdd370a6c72a4ef4641c4738cf31b1" exitCode=0 Dec 05 10:47:20.360975 master-0 kubenswrapper[24928]: I1205 10:47:20.360937 24928 generic.go:334] "Generic (PLEG): container finished" podID="d95a56ba-c940-4e3e-aed6-d8c04f1871b6" containerID="dad5a560aae0ea7cbefd19607cf4263a0d5cfc63a06ebe4360483bee7047a720" exitCode=0 Dec 05 10:47:20.363349 master-0 kubenswrapper[24928]: I1205 10:47:20.363295 24928 generic.go:334] "Generic (PLEG): container finished" podID="926263c4-ec5b-41cb-9c30-0c88f636035f" containerID="3504f06691ba2b115136a2091ae95118383b7c2cc711e4550b1e60138be21510" exitCode=0 Dec 05 10:47:20.369183 master-0 kubenswrapper[24928]: I1205 10:47:20.369135 24928 generic.go:334] "Generic (PLEG): container finished" podID="baa812e4-b011-41ee-82ab-8f571f0b7e0a" containerID="84efa2150952ddba357e673dac2d7ef63c7f6bced1cf84122ed7eed4c4e74f80" exitCode=0 Dec 05 10:47:20.369183 master-0 kubenswrapper[24928]: I1205 10:47:20.369182 24928 generic.go:334] "Generic (PLEG): container finished" podID="baa812e4-b011-41ee-82ab-8f571f0b7e0a" containerID="b0504c32d8948d61de04b9482cfad535351eb3152188b508b07ba9be49d98d54" exitCode=0 Dec 05 10:47:20.371977 master-0 kubenswrapper[24928]: I1205 10:47:20.371927 24928 generic.go:334] "Generic (PLEG): container finished" podID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerID="e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23" exitCode=0 Dec 05 10:47:20.372051 master-0 kubenswrapper[24928]: I1205 10:47:20.371978 24928 generic.go:334] "Generic (PLEG): container finished" podID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerID="698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90" exitCode=0 Dec 05 10:47:20.373697 master-0 kubenswrapper[24928]: I1205 10:47:20.373670 24928 generic.go:334] "Generic (PLEG): container finished" podID="70fbf150-c1f7-4857-8b63-c566282e2526" containerID="2215577fb4531349eb7213f364526a3db417316312ccf7c18444e96dbaf859c1" exitCode=0 Dec 05 10:47:20.375674 master-0 kubenswrapper[24928]: I1205 10:47:20.375624 24928 generic.go:334] "Generic (PLEG): container finished" podID="fd58232c-a81a-4aee-8b2c-5ffcdded2e23" containerID="3908741f62d9038e3367c49d1d01f504f11849e5b644100de90c2c7d751cc2e3" exitCode=0 Dec 05 10:47:20.386467 master-0 kubenswrapper[24928]: I1205 10:47:20.385871 24928 generic.go:334] "Generic (PLEG): container finished" podID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerID="8aa1ac52d0f6c8b159f0da42738bae856c9e11f506fdfed545be3d79a9c5ba1d" exitCode=0 Dec 05 10:47:20.391014 master-0 kubenswrapper[24928]: I1205 10:47:20.390985 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-config-operator_kube-rbac-proxy-crio-master-0_3169f44496ed8a28c6d6a15511ab0eec/kube-rbac-proxy-crio/2.log" Dec 05 10:47:20.391579 master-0 kubenswrapper[24928]: I1205 10:47:20.391531 24928 generic.go:334] "Generic (PLEG): container finished" podID="3169f44496ed8a28c6d6a15511ab0eec" containerID="c4c2663f7fd33f4762e4466ae563543a600762863049b80576e8bb073c24bbd0" exitCode=1 Dec 05 10:47:20.391579 master-0 kubenswrapper[24928]: I1205 10:47:20.391575 24928 generic.go:334] "Generic (PLEG): container finished" podID="3169f44496ed8a28c6d6a15511ab0eec" containerID="2a995b4c36846fd814a8f2ca3bd01eb60eaabd9a004a171a036b25b030f63eff" exitCode=0 Dec 05 10:47:20.393374 master-0 kubenswrapper[24928]: I1205 10:47:20.393347 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-2-master-0_68058d25-5888-495f-8571-453f0bf919d9/installer/0.log" Dec 05 10:47:20.393459 master-0 kubenswrapper[24928]: I1205 10:47:20.393399 24928 generic.go:334] "Generic (PLEG): container finished" podID="68058d25-5888-495f-8571-453f0bf919d9" containerID="2ea94dd01104138f23272416f98911388c3b177dca6ad44593587087deb08528" exitCode=1 Dec 05 10:47:20.396702 master-0 kubenswrapper[24928]: I1205 10:47:20.396675 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/0.log" Dec 05 10:47:20.396754 master-0 kubenswrapper[24928]: I1205 10:47:20.396707 24928 generic.go:334] "Generic (PLEG): container finished" podID="e27c0798-ec1c-43cd-b81b-f77f2f11ad0f" containerID="4308a58237214eca1f5ba917e53d41b87b4e41bf19760a2305217ef9ccb8ed3f" exitCode=1 Dec 05 10:47:20.403317 master-0 kubenswrapper[24928]: I1205 10:47:20.403232 24928 generic.go:334] "Generic (PLEG): container finished" podID="b89698aa356a3bc32694e2b098f9a900" containerID="c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b" exitCode=0 Dec 05 10:47:20.410357 master-0 kubenswrapper[24928]: I1205 10:47:20.410310 24928 generic.go:334] "Generic (PLEG): container finished" podID="b760849c-8d83-47da-8677-68445c143bef" containerID="d5454b7ce1bc247671ab46448edd5a6f0a198a1673e48f822fea3525f6db868a" exitCode=0 Dec 05 10:47:20.419204 master-0 kubenswrapper[24928]: I1205 10:47:20.419166 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-controller_operator-controller-controller-manager-7cbd59c7f8-dh5tt_592373ae-a952-4dd3-a9bc-f9c9c19c0802/manager/0.log" Dec 05 10:47:20.419383 master-0 kubenswrapper[24928]: I1205 10:47:20.419229 24928 generic.go:334] "Generic (PLEG): container finished" podID="592373ae-a952-4dd3-a9bc-f9c9c19c0802" containerID="665b41c4b05327dcf1d9daca6aa21544857362b43e6c7d8ef7fac95d7153f9b2" exitCode=1 Dec 05 10:47:20.443291 master-0 kubenswrapper[24928]: I1205 10:47:20.443227 24928 generic.go:334] "Generic (PLEG): container finished" podID="f1d74f88-1419-431e-80da-26db419f050e" containerID="afde21a1bd7da8808580dd974998cee92cb64026aa792ec094def214f272bc2f" exitCode=0 Dec 05 10:47:20.447599 master-0 kubenswrapper[24928]: I1205 10:47:20.447534 24928 generic.go:334] "Generic (PLEG): container finished" podID="445d75af-d072-4fa0-91a7-f3fa579b9ca9" containerID="df3d2e1cdc52294902ada6e0e872265f937f9cd41ea5e7643f484cd5a4774ce3" exitCode=0 Dec 05 10:47:20.449100 master-0 kubenswrapper[24928]: I1205 10:47:20.449058 24928 generic.go:334] "Generic (PLEG): container finished" podID="75ba20a1-b322-4ce8-88d2-d8ee210cc27a" containerID="c8f54c8d7593dfbeebbefc04c820ade3fd814ef907ea5e95f2df34b8ed43589a" exitCode=0 Dec 05 10:47:20.452594 master-0 kubenswrapper[24928]: I1205 10:47:20.452564 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/3.log" Dec 05 10:47:20.453657 master-0 kubenswrapper[24928]: I1205 10:47:20.453613 24928 generic.go:334] "Generic (PLEG): container finished" podID="22676fac-b770-4937-9bee-7478bd1babb7" containerID="76b36235b0c1dc04c6934f55253404fa8e8aa924bd0ff655fd03d44b54d22caa" exitCode=1 Dec 05 10:47:20.456678 master-0 kubenswrapper[24928]: I1205 10:47:20.456645 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_installer-1-master-0_738d9fe7-2d08-4aa4-842b-9cbeddf120fb/installer/0.log" Dec 05 10:47:20.456756 master-0 kubenswrapper[24928]: I1205 10:47:20.456710 24928 generic.go:334] "Generic (PLEG): container finished" podID="738d9fe7-2d08-4aa4-842b-9cbeddf120fb" containerID="8a8e52c1bfb7c50d31dbfbb099e72a94f8cf7f5c9eac30d42a9e7bb6f30721bf" exitCode=1 Dec 05 10:47:20.503253 master-0 kubenswrapper[24928]: E1205 10:47:20.503201 24928 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 05 10:47:20.602949 master-0 kubenswrapper[24928]: I1205 10:47:20.602890 24928 manager.go:324] Recovery completed Dec 05 10:47:20.691336 master-0 kubenswrapper[24928]: I1205 10:47:20.691184 24928 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 05 10:47:20.691336 master-0 kubenswrapper[24928]: I1205 10:47:20.691218 24928 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 05 10:47:20.691336 master-0 kubenswrapper[24928]: I1205 10:47:20.691250 24928 state_mem.go:36] "Initialized new in-memory state store" Dec 05 10:47:20.691601 master-0 kubenswrapper[24928]: I1205 10:47:20.691443 24928 state_mem.go:88] "Updated default CPUSet" cpuSet="" Dec 05 10:47:20.691601 master-0 kubenswrapper[24928]: I1205 10:47:20.691461 24928 state_mem.go:96] "Updated CPUSet assignments" assignments={} Dec 05 10:47:20.691601 master-0 kubenswrapper[24928]: I1205 10:47:20.691485 24928 state_checkpoint.go:136] "State checkpoint: restored state from checkpoint" Dec 05 10:47:20.691601 master-0 kubenswrapper[24928]: I1205 10:47:20.691493 24928 state_checkpoint.go:137] "State checkpoint: defaultCPUSet" defaultCpuSet="" Dec 05 10:47:20.691601 master-0 kubenswrapper[24928]: I1205 10:47:20.691501 24928 policy_none.go:49] "None policy: Start" Dec 05 10:47:20.697841 master-0 kubenswrapper[24928]: I1205 10:47:20.697783 24928 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 05 10:47:20.697928 master-0 kubenswrapper[24928]: I1205 10:47:20.697854 24928 state_mem.go:35] "Initializing new in-memory state store" Dec 05 10:47:20.698150 master-0 kubenswrapper[24928]: I1205 10:47:20.698118 24928 state_mem.go:75] "Updated machine memory state" Dec 05 10:47:20.698150 master-0 kubenswrapper[24928]: I1205 10:47:20.698139 24928 state_checkpoint.go:82] "State checkpoint: restored state from checkpoint" Dec 05 10:47:20.715994 master-0 kubenswrapper[24928]: I1205 10:47:20.715920 24928 manager.go:334] "Starting Device Plugin manager" Dec 05 10:47:20.716214 master-0 kubenswrapper[24928]: I1205 10:47:20.716073 24928 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 05 10:47:20.716214 master-0 kubenswrapper[24928]: I1205 10:47:20.716091 24928 server.go:79] "Starting device plugin registration server" Dec 05 10:47:20.716589 master-0 kubenswrapper[24928]: I1205 10:47:20.716558 24928 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 05 10:47:20.716652 master-0 kubenswrapper[24928]: I1205 10:47:20.716579 24928 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 05 10:47:20.716796 master-0 kubenswrapper[24928]: I1205 10:47:20.716743 24928 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 05 10:47:20.716976 master-0 kubenswrapper[24928]: I1205 10:47:20.716944 24928 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 05 10:47:20.716976 master-0 kubenswrapper[24928]: I1205 10:47:20.716968 24928 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 05 10:47:20.817445 master-0 kubenswrapper[24928]: I1205 10:47:20.817380 24928 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:47:20.822386 master-0 kubenswrapper[24928]: I1205 10:47:20.822327 24928 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:47:20.822386 master-0 kubenswrapper[24928]: I1205 10:47:20.822373 24928 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:47:20.822386 master-0 kubenswrapper[24928]: I1205 10:47:20.822383 24928 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:47:20.822715 master-0 kubenswrapper[24928]: I1205 10:47:20.822514 24928 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:47:20.825896 master-0 kubenswrapper[24928]: E1205 10:47:20.825851 24928 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"master-0\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="master-0" Dec 05 10:47:20.904372 master-0 kubenswrapper[24928]: I1205 10:47:20.904236 24928 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-master-0","openshift-etcd/etcd-master-0","openshift-kube-apiserver/kube-apiserver-master-0","openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0","openshift-kube-controller-manager/kube-controller-manager-master-0","openshift-kube-scheduler/openshift-kube-scheduler-master-0"] Dec 05 10:47:20.905470 master-0 kubenswrapper[24928]: I1205 10:47:20.905262 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4293a095feecc2a652807fec9aa96d59e590e18262f101aef59da2ccb3b4fbf6" Dec 05 10:47:20.905470 master-0 kubenswrapper[24928]: I1205 10:47:20.905332 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a87c5cb2ce0ea32814514a0a2632b27b31a950e85396e1a0393057f93ec2e154" Dec 05 10:47:20.905470 master-0 kubenswrapper[24928]: I1205 10:47:20.905342 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adcd1c5ddd833ada6b59dc1ab4b6b3a5842f80f520ee65ce65a40c8afe637eab" Dec 05 10:47:20.905470 master-0 kubenswrapper[24928]: I1205 10:47:20.905359 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0127ee22dc31281cc8a51e81727d0320a63ff8209b1fadb0e73c9f87ae0d74aa" Dec 05 10:47:20.905470 master-0 kubenswrapper[24928]: I1205 10:47:20.905383 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819"} Dec 05 10:47:20.905470 master-0 kubenswrapper[24928]: I1205 10:47:20.905452 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526"} Dec 05 10:47:20.905470 master-0 kubenswrapper[24928]: I1205 10:47:20.905462 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5"} Dec 05 10:47:20.905470 master-0 kubenswrapper[24928]: I1205 10:47:20.905470 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e"} Dec 05 10:47:20.905470 master-0 kubenswrapper[24928]: I1205 10:47:20.905478 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab"} Dec 05 10:47:20.905470 master-0 kubenswrapper[24928]: I1205 10:47:20.905487 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerDied","Data":"41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905498 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerDied","Data":"d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905506 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerDied","Data":"f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905515 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"c24e01603234fe8003f8aae8171b0065","Type":"ContainerStarted","Data":"9e8deb0a79a04a088af836af85742626a8778be7f61d22448e3a0b846ff8b05a"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905542 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerStarted","Data":"29ac63e75838b26e5ffeb47eea93679b6f563ae1d361a9d774732d46e7d9c13d"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905551 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerStarted","Data":"804cb8acb4e8cb78d7a134f3b11dcb55d43d2e3b4630654676497a1d400dbc6f"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905560 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerStarted","Data":"8c491ff443ec7786bd22bb8e647eeff7f1b6e812a627f82eec65955c76037686"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905569 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerDied","Data":"baa0fb92990e2611d666dd15d06398cf84d61306c4b64819cbec0b7a2dd1f1f8"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905578 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerStarted","Data":"b73ea430e90c45d058f4dc00766a5e44931feca970feb8976e3fca4580bc112f"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905603 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="690adabd380efbcf30cb261955bc2a3bb79595c5d056733e17f611fb3912f863" Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905643 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba7c2d250edf116aab8b9c6bc4e16427c460bc08a2f3f72ce479a134eaf54756" Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905650 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"f7c86a3fbacaa184c777e4b5c80c29c00ccd9777b6aeed8a46d62880a41cc869"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905658 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"47bef9607325414e871a594b06d4250139a33f93aa8b2368ce9838e31d093051"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905667 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"00db1aab947f0a6caa3edb56d388eb498e266a62880572bf52a5608eb626a365"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905675 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"a0c4bf77b56c9bf53793c3092e77173ade30a59f4de69720ced91f80c7a365c6"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905683 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"5be78d3569389b67ef6d6c6575e7c97a75617917f6e7f31b9b05a692b9e12292"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905699 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="512588b6c638c1ce7fb1dce16bbe4f74e94e57511137324c00c995274f332c5c" Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905731 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="452fb60a08a0303dfa4a44b4cff2ac5af4105f5fed0d2d7d9442e582b765e031" Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905753 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerStarted","Data":"3a88ca4d52cee88aa8dcb914ea7d8cff8936f823056376f936c3da2aa8e9a481"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905762 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerDied","Data":"c4c2663f7fd33f4762e4466ae563543a600762863049b80576e8bb073c24bbd0"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905773 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerDied","Data":"2a995b4c36846fd814a8f2ca3bd01eb60eaabd9a004a171a036b25b030f63eff"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905785 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" event={"ID":"3169f44496ed8a28c6d6a15511ab0eec","Type":"ContainerStarted","Data":"d14f8103147de11952469f9448b1a161dedb77f6737557ed8a9f9a9dd87b69e7"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905799 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7533cfa355ed03950ce16c744c7022b947100926d9199c99ab99b289da6cb649" Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905814 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" event={"ID":"a906debd0c35952850935aee2d607cce","Type":"ContainerStarted","Data":"8930f7af011c1e43a64ba6a4ca85659538a4774816c777a7c34e7a4b10c3cab3"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905826 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" event={"ID":"a906debd0c35952850935aee2d607cce","Type":"ContainerStarted","Data":"4446e01522dc26b79f498f8e0dc137457645ab081cac6529b44e9b557b9d72f8"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905844 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905856 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905869 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905881 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905892 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905905 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerDied","Data":"c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b"} Dec 05 10:47:20.905904 master-0 kubenswrapper[24928]: I1205 10:47:20.905919 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"b89698aa356a3bc32694e2b098f9a900","Type":"ContainerStarted","Data":"b94fba3ebe8eacc577514add51b89ec37946322bc96bbe0ee6d3e8d5b7830ae3"} Dec 05 10:47:20.906735 master-0 kubenswrapper[24928]: I1205 10:47:20.905949 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09d6c752fa5b21d6de28695ca1e30252206036602e7b7cad7c82ecf0c820105d" Dec 05 10:47:20.906735 master-0 kubenswrapper[24928]: I1205 10:47:20.906001 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e72f14b85f51789b3236420aa213917fe54cddf2ceedaab3f6bad48e1ce12513" Dec 05 10:47:20.906735 master-0 kubenswrapper[24928]: I1205 10:47:20.906024 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b1988a80333a5d160a32b470d3b1aa05cd5259d79366c745b970eff4d6866a0" Dec 05 10:47:20.906735 master-0 kubenswrapper[24928]: I1205 10:47:20.906043 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="481bec329814be919f1e3dd3cd8fba26d8d823339446a0f9b66a76525352b23f" Dec 05 10:47:20.916076 master-0 kubenswrapper[24928]: E1205 10:47:20.915998 24928 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-master-0\" already exists" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:20.925590 master-0 kubenswrapper[24928]: E1205 10:47:20.925513 24928 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-master-0\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:20.928004 master-0 kubenswrapper[24928]: E1205 10:47:20.927944 24928 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-master-0\" already exists" pod="openshift-etcd/etcd-master-0" Dec 05 10:47:20.928004 master-0 kubenswrapper[24928]: E1205 10:47:20.927979 24928 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"openshift-kube-scheduler-master-0\" already exists" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:47:20.977616 master-0 kubenswrapper[24928]: I1205 10:47:20.977405 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-data-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:20.977616 master-0 kubenswrapper[24928]: I1205 10:47:20.977494 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-cert-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:20.977616 master-0 kubenswrapper[24928]: I1205 10:47:20.977518 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-audit-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:20.977616 master-0 kubenswrapper[24928]: I1205 10:47:20.977537 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-resource-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:47:20.977616 master-0 kubenswrapper[24928]: I1205 10:47:20.977554 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-etc-kube\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:47:20.977996 master-0 kubenswrapper[24928]: I1205 10:47:20.977692 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-usr-local-bin\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:20.977996 master-0 kubenswrapper[24928]: I1205 10:47:20.977861 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-log-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:20.977996 master-0 kubenswrapper[24928]: I1205 10:47:20.977927 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-log\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:20.977996 master-0 kubenswrapper[24928]: I1205 10:47:20.977971 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:20.978259 master-0 kubenswrapper[24928]: I1205 10:47:20.978004 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:47:20.978259 master-0 kubenswrapper[24928]: I1205 10:47:20.978038 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-cert-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:20.978259 master-0 kubenswrapper[24928]: I1205 10:47:20.978070 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-static-pod-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:20.978259 master-0 kubenswrapper[24928]: I1205 10:47:20.978100 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-resource-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:20.978259 master-0 kubenswrapper[24928]: I1205 10:47:20.978217 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-resource-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:20.978577 master-0 kubenswrapper[24928]: I1205 10:47:20.978258 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:20.978577 master-0 kubenswrapper[24928]: I1205 10:47:20.978297 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:20.978577 master-0 kubenswrapper[24928]: I1205 10:47:20.978337 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-cert-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:47:20.978577 master-0 kubenswrapper[24928]: I1205 10:47:20.978383 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:20.978577 master-0 kubenswrapper[24928]: I1205 10:47:20.978403 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-manifests\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:20.978577 master-0 kubenswrapper[24928]: I1205 10:47:20.978442 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-lock\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:21.026628 master-0 kubenswrapper[24928]: I1205 10:47:21.026522 24928 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:47:21.030150 master-0 kubenswrapper[24928]: I1205 10:47:21.030109 24928 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:47:21.030613 master-0 kubenswrapper[24928]: I1205 10:47:21.030160 24928 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:47:21.030613 master-0 kubenswrapper[24928]: I1205 10:47:21.030173 24928 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:47:21.030613 master-0 kubenswrapper[24928]: I1205 10:47:21.030295 24928 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:47:21.035277 master-0 kubenswrapper[24928]: E1205 10:47:21.035225 24928 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"master-0\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="master-0" Dec 05 10:47:21.079047 master-0 kubenswrapper[24928]: I1205 10:47:21.078953 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-cert-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:21.079047 master-0 kubenswrapper[24928]: I1205 10:47:21.079015 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-audit-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079109 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-cert-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079171 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-resource-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079216 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-etc-kube\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079220 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-audit-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079242 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-resource-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079252 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-data-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079306 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-data-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079320 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-log-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079367 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-log-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079369 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-log\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079401 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-log\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079445 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-etc-kube\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079409 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079468 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079496 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079517 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-cert-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079535 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-usr-local-bin\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079552 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-resource-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.079537 master-0 kubenswrapper[24928]: I1205 10:47:21.079569 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-resource-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079686 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-cert-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079746 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079770 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-resource-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079781 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-usr-local-bin\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079787 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079830 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079830 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079870 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-cert-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079882 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-resource-dir\") pod \"kube-apiserver-master-0\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079796 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3169f44496ed8a28c6d6a15511ab0eec-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-master-0\" (UID: \"3169f44496ed8a28c6d6a15511ab0eec\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079900 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-static-pod-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079909 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/bdd20c68f1ccfd1240fb273af0638d01-cert-dir\") pod \"openshift-kube-scheduler-master-0\" (UID: \"bdd20c68f1ccfd1240fb273af0638d01\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079929 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-static-pod-dir\") pod \"etcd-master-0\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079930 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079975 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-manifests\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.079996 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-lock\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.080035 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-lock\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.080066 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:21.081680 master-0 kubenswrapper[24928]: I1205 10:47:21.080070 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-manifests\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"a906debd0c35952850935aee2d607cce\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:21.120253 master-0 kubenswrapper[24928]: I1205 10:47:21.120167 24928 apiserver.go:52] "Watching apiserver" Dec 05 10:47:21.146464 master-0 kubenswrapper[24928]: I1205 10:47:21.146308 24928 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 10:47:21.149266 master-0 kubenswrapper[24928]: I1205 10:47:21.149147 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-lxmgz","openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c","openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f","openshift-insights/insights-operator-55965856b6-2sxv7","openshift-kube-controller-manager/installer-1-master-0","openshift-kube-controller-manager/installer-2-master-0","openshift-marketplace/redhat-operators-pqhfn","openshift-cluster-node-tuning-operator/tuned-hvh88","openshift-marketplace/marketplace-operator-f797b99b6-z9qcl","openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn","openshift-controller-manager/controller-manager-86f4478dbf-jqlt9","openshift-kube-scheduler/installer-4-master-0","openshift-marketplace/certified-operators-djhk8","openshift-monitoring/metrics-server-7c46d76dff-z8d8z","openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77","openshift-apiserver/apiserver-5b9fd577f8-6sxcx","openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm","openshift-multus/multus-additional-cni-plugins-dms5d","openshift-machine-api/machine-api-operator-88d48b57d-x7jfs","openshift-marketplace/community-operators-6p8cq","openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8","openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87","openshift-etcd/installer-2-master-0","openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb","openshift-kube-apiserver/installer-3-master-0","openshift-etcd/installer-1-master-0","openshift-kube-controller-manager/kube-controller-manager-master-0","openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs","openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz","openshift-ovn-kubernetes/ovnkube-node-rsfjs","openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t","openshift-kube-scheduler/installer-3-master-0","openshift-machine-config-operator/machine-config-server-5t4nn","openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps","openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5","openshift-etcd/etcd-master-0","openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8","openshift-kube-controller-manager/installer-3-master-0","openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp","openshift-ingress-canary/ingress-canary-knq92","openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t","openshift-monitoring/telemeter-client-86cb595668-52qnw","assisted-installer/assisted-installer-controller-pd4q6","openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27","openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z","openshift-oauth-apiserver/apiserver-85b8f855df-8g52w","openshift-service-ca/service-ca-77c99c46b8-m7zqs","openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s","openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4","openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv","openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h","openshift-monitoring/node-exporter-bmqsb","openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt","openshift-multus/network-metrics-daemon-8gjgm","openshift-network-node-identity/network-node-identity-ql7j7","openshift-network-operator/network-operator-79767b7ff9-t8j2j","openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7","openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p","openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v","openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f","openshift-marketplace/redhat-marketplace-wk29h","openshift-monitoring/kube-state-metrics-5857974f64-xj7pj","openshift-network-diagnostics/network-check-target-d6fzk","openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk","openshift-ingress-operator/ingress-operator-8649c48786-cgt5x","openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl","openshift-kube-apiserver/installer-1-master-0","openshift-kube-scheduler/openshift-kube-scheduler-master-0","openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz","openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c","openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr","openshift-multus/multus-admission-controller-8dbbb5754-7p9c2","openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk","openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6","openshift-dns/node-resolver-qkccw","openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0","openshift-machine-config-operator/kube-rbac-proxy-crio-master-0","openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h","openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw","openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt","openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7","openshift-dns/dns-default-4vxng","openshift-ingress/router-default-5465c8b4db-s4c2f","openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd","openshift-machine-config-operator/machine-config-daemon-5n6nw","openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7","openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p","openshift-kube-apiserver/kube-apiserver-master-0","openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr","openshift-network-operator/iptables-alerter-d6wjk"] Dec 05 10:47:21.149649 master-0 kubenswrapper[24928]: I1205 10:47:21.149577 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="assisted-installer/assisted-installer-controller-pd4q6" Dec 05 10:47:21.153275 master-0 kubenswrapper[24928]: I1205 10:47:21.153021 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 10:47:21.154838 master-0 kubenswrapper[24928]: I1205 10:47:21.154802 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 10:47:21.155898 master-0 kubenswrapper[24928]: I1205 10:47:21.155781 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 10:47:21.156016 master-0 kubenswrapper[24928]: I1205 10:47:21.155953 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 10:47:21.158796 master-0 kubenswrapper[24928]: I1205 10:47:21.158747 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 10:47:21.158912 master-0 kubenswrapper[24928]: I1205 10:47:21.158850 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Dec 05 10:47:21.160847 master-0 kubenswrapper[24928]: I1205 10:47:21.160810 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.161024 master-0 kubenswrapper[24928]: I1205 10:47:21.160977 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.161118 master-0 kubenswrapper[24928]: I1205 10:47:21.161070 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 10:47:21.161118 master-0 kubenswrapper[24928]: I1205 10:47:21.160888 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 10:47:21.161347 master-0 kubenswrapper[24928]: I1205 10:47:21.161312 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 10:47:21.161529 master-0 kubenswrapper[24928]: I1205 10:47:21.161370 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 10:47:21.161529 master-0 kubenswrapper[24928]: I1205 10:47:21.161314 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 10:47:21.161529 master-0 kubenswrapper[24928]: I1205 10:47:21.161486 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.161529 master-0 kubenswrapper[24928]: I1205 10:47:21.161521 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 10:47:21.161779 master-0 kubenswrapper[24928]: I1205 10:47:21.161627 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Dec 05 10:47:21.161779 master-0 kubenswrapper[24928]: I1205 10:47:21.161642 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.162195 master-0 kubenswrapper[24928]: I1205 10:47:21.162155 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 10:47:21.162335 master-0 kubenswrapper[24928]: I1205 10:47:21.162304 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-1-master-0" Dec 05 10:47:21.162416 master-0 kubenswrapper[24928]: I1205 10:47:21.162347 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 10:47:21.162725 master-0 kubenswrapper[24928]: I1205 10:47:21.162694 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 10:47:21.162725 master-0 kubenswrapper[24928]: I1205 10:47:21.162710 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Dec 05 10:47:21.162879 master-0 kubenswrapper[24928]: I1205 10:47:21.162804 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Dec 05 10:47:21.162951 master-0 kubenswrapper[24928]: I1205 10:47:21.162938 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 10:47:21.164877 master-0 kubenswrapper[24928]: I1205 10:47:21.164800 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 10:47:21.165991 master-0 kubenswrapper[24928]: I1205 10:47:21.165771 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 10:47:21.166356 master-0 kubenswrapper[24928]: I1205 10:47:21.166310 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 10:47:21.166356 master-0 kubenswrapper[24928]: I1205 10:47:21.166337 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 10:47:21.166508 master-0 kubenswrapper[24928]: I1205 10:47:21.166385 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 10:47:21.166786 master-0 kubenswrapper[24928]: I1205 10:47:21.166754 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.166870 master-0 kubenswrapper[24928]: I1205 10:47:21.166805 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 10:47:21.166946 master-0 kubenswrapper[24928]: I1205 10:47:21.166836 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 10:47:21.167013 master-0 kubenswrapper[24928]: I1205 10:47:21.166962 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 10:47:21.167162 master-0 kubenswrapper[24928]: I1205 10:47:21.167132 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-storage-operator"/"kube-root-ca.crt" Dec 05 10:47:21.169106 master-0 kubenswrapper[24928]: I1205 10:47:21.169071 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 10:47:21.169248 master-0 kubenswrapper[24928]: I1205 10:47:21.169208 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-storage-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.169398 master-0 kubenswrapper[24928]: I1205 10:47:21.169366 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 10:47:21.169489 master-0 kubenswrapper[24928]: I1205 10:47:21.169395 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 10:47:21.170007 master-0 kubenswrapper[24928]: I1205 10:47:21.169977 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-olm-operator"/"cluster-olm-operator-serving-cert" Dec 05 10:47:21.170213 master-0 kubenswrapper[24928]: I1205 10:47:21.170188 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:47:21.170369 master-0 kubenswrapper[24928]: I1205 10:47:21.170338 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.171793 master-0 kubenswrapper[24928]: I1205 10:47:21.171079 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 10:47:21.171793 master-0 kubenswrapper[24928]: I1205 10:47:21.171212 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-olm-operator"/"kube-root-ca.crt" Dec 05 10:47:21.174842 master-0 kubenswrapper[24928]: I1205 10:47:21.172877 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-olm-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.174842 master-0 kubenswrapper[24928]: I1205 10:47:21.173089 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.174842 master-0 kubenswrapper[24928]: I1205 10:47:21.173112 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 10:47:21.174842 master-0 kubenswrapper[24928]: I1205 10:47:21.173636 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 10:47:21.174842 master-0 kubenswrapper[24928]: I1205 10:47:21.173709 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 10:47:21.175134 master-0 kubenswrapper[24928]: I1205 10:47:21.175051 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.175331 master-0 kubenswrapper[24928]: I1205 10:47:21.175229 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 10:47:21.175489 master-0 kubenswrapper[24928]: I1205 10:47:21.175245 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 10:47:21.175489 master-0 kubenswrapper[24928]: I1205 10:47:21.175249 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 10:47:21.175624 master-0 kubenswrapper[24928]: I1205 10:47:21.175504 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.175624 master-0 kubenswrapper[24928]: I1205 10:47:21.175405 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 10:47:21.176384 master-0 kubenswrapper[24928]: I1205 10:47:21.176338 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-1-master-0" Dec 05 10:47:21.176717 master-0 kubenswrapper[24928]: I1205 10:47:21.176671 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-3-master-0" Dec 05 10:47:21.177766 master-0 kubenswrapper[24928]: I1205 10:47:21.177720 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 10:47:21.177864 master-0 kubenswrapper[24928]: I1205 10:47:21.177839 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 10:47:21.178447 master-0 kubenswrapper[24928]: I1205 10:47:21.178380 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 10:47:21.184455 master-0 kubenswrapper[24928]: I1205 10:47:21.184368 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 10:47:21.184773 master-0 kubenswrapper[24928]: I1205 10:47:21.184720 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 10:47:21.184866 master-0 kubenswrapper[24928]: I1205 10:47:21.184783 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 10:47:21.184866 master-0 kubenswrapper[24928]: I1205 10:47:21.184812 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 10:47:21.184866 master-0 kubenswrapper[24928]: I1205 10:47:21.184843 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 10:47:21.185057 master-0 kubenswrapper[24928]: I1205 10:47:21.184865 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"whereabouts-config" Dec 05 10:47:21.185057 master-0 kubenswrapper[24928]: I1205 10:47:21.184893 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 10:47:21.185057 master-0 kubenswrapper[24928]: I1205 10:47:21.185034 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"kube-root-ca.crt" Dec 05 10:47:21.185513 master-0 kubenswrapper[24928]: I1205 10:47:21.185462 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.185699 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.185749 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.185912 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.185968 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.185985 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"openshift-service-ca.crt" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.186067 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-node-tuning-operator"/"performance-addon-operator-webhook-cert" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.186120 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.186150 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-node-tuning-operator"/"node-tuning-operator-tls" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.186269 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.186327 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.186322 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.186457 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.186484 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.186544 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.186653 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 10:47:21.187183 master-0 kubenswrapper[24928]: I1205 10:47:21.186767 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 10:47:21.188087 master-0 kubenswrapper[24928]: I1205 10:47:21.187354 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 10:47:21.188087 master-0 kubenswrapper[24928]: I1205 10:47:21.187440 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:47:21.188323 master-0 kubenswrapper[24928]: I1205 10:47:21.188252 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 10:47:21.189200 master-0 kubenswrapper[24928]: I1205 10:47:21.188971 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-1-master-0" Dec 05 10:47:21.190593 master-0 kubenswrapper[24928]: I1205 10:47:21.189335 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 10:47:21.190593 master-0 kubenswrapper[24928]: I1205 10:47:21.189827 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 10:47:21.190593 master-0 kubenswrapper[24928]: I1205 10:47:21.189908 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 10:47:21.190593 master-0 kubenswrapper[24928]: I1205 10:47:21.190339 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 10:47:21.190593 master-0 kubenswrapper[24928]: I1205 10:47:21.190516 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 10:47:21.191100 master-0 kubenswrapper[24928]: I1205 10:47:21.190822 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-2-master-0" Dec 05 10:47:21.193777 master-0 kubenswrapper[24928]: I1205 10:47:21.193703 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 10:47:21.193892 master-0 kubenswrapper[24928]: I1205 10:47:21.193718 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 10:47:21.193892 master-0 kubenswrapper[24928]: I1205 10:47:21.193851 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 10:47:21.195245 master-0 kubenswrapper[24928]: I1205 10:47:21.195161 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 10:47:21.195490 master-0 kubenswrapper[24928]: I1205 10:47:21.195295 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 10:47:21.197949 master-0 kubenswrapper[24928]: I1205 10:47:21.197909 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 10:47:21.198151 master-0 kubenswrapper[24928]: I1205 10:47:21.198002 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 10:47:21.201956 master-0 kubenswrapper[24928]: I1205 10:47:21.201922 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"trusted-ca" Dec 05 10:47:21.203724 master-0 kubenswrapper[24928]: I1205 10:47:21.203683 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 10:47:21.203802 master-0 kubenswrapper[24928]: I1205 10:47:21.203741 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-3-master-0" Dec 05 10:47:21.204242 master-0 kubenswrapper[24928]: I1205 10:47:21.204213 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7" Dec 05 10:47:21.204336 master-0 kubenswrapper[24928]: I1205 10:47:21.204220 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/installer-4-master-0" Dec 05 10:47:21.205512 master-0 kubenswrapper[24928]: I1205 10:47:21.205279 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:21.205512 master-0 kubenswrapper[24928]: I1205 10:47:21.205369 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:21.210475 master-0 kubenswrapper[24928]: I1205 10:47:21.209797 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 10:47:21.213210 master-0 kubenswrapper[24928]: I1205 10:47:21.212677 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 10:47:21.215725 master-0 kubenswrapper[24928]: I1205 10:47:21.215634 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 10:47:21.217609 master-0 kubenswrapper[24928]: I1205 10:47:21.217546 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 10:47:21.232846 master-0 kubenswrapper[24928]: I1205 10:47:21.232735 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 10:47:21.253080 master-0 kubenswrapper[24928]: I1205 10:47:21.253033 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 10:47:21.253545 master-0 kubenswrapper[24928]: I1205 10:47:21.253482 24928 desired_state_of_world_populator.go:155] "Finished populating initial desired state of world" Dec 05 10:47:21.273343 master-0 kubenswrapper[24928]: I1205 10:47:21.273270 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 10:47:21.285173 master-0 kubenswrapper[24928]: I1205 10:47:21.285096 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-node-bootstrap-token\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:47:21.285414 master-0 kubenswrapper[24928]: I1205 10:47:21.285298 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-metrics-client-ca\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:47:21.285414 master-0 kubenswrapper[24928]: I1205 10:47:21.285348 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-images\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:21.285414 master-0 kubenswrapper[24928]: I1205 10:47:21.285374 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-auth-proxy-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:47:21.285414 master-0 kubenswrapper[24928]: I1205 10:47:21.285406 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-hostroot\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.285584 master-0 kubenswrapper[24928]: I1205 10:47:21.285453 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-config\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:21.285584 master-0 kubenswrapper[24928]: I1205 10:47:21.285487 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-auth-proxy-config\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:21.285584 master-0 kubenswrapper[24928]: I1205 10:47:21.285516 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-env-overrides\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:47:21.285584 master-0 kubenswrapper[24928]: I1205 10:47:21.285544 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:47:21.285877 master-0 kubenswrapper[24928]: I1205 10:47:21.285821 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:21.285877 master-0 kubenswrapper[24928]: I1205 10:47:21.285856 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-config\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:21.286043 master-0 kubenswrapper[24928]: I1205 10:47:21.285881 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-lib-modules\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.286043 master-0 kubenswrapper[24928]: I1205 10:47:21.285910 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-multus-daemon-config\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.286043 master-0 kubenswrapper[24928]: I1205 10:47:21.285931 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-metrics-client-ca\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:21.286043 master-0 kubenswrapper[24928]: I1205 10:47:21.285941 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-env-overrides\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:47:21.286043 master-0 kubenswrapper[24928]: I1205 10:47:21.285952 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-trusted-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:21.286043 master-0 kubenswrapper[24928]: I1205 10:47:21.285997 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m88gw\" (UniqueName: \"kubernetes.io/projected/d5198438-06ae-4e63-a7e3-950ba23bba9c-kube-api-access-m88gw\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.286043 master-0 kubenswrapper[24928]: I1205 10:47:21.286028 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dtfn\" (UniqueName: \"kubernetes.io/projected/5a1bdc70-6412-47e0-8330-04d796cc8d55-kube-api-access-5dtfn\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:47:21.286310 master-0 kubenswrapper[24928]: I1205 10:47:21.286057 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-systemd-units\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.286310 master-0 kubenswrapper[24928]: I1205 10:47:21.286088 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f69hj\" (UniqueName: \"kubernetes.io/projected/48bd1d86-a6f2-439f-ab04-6a9a442bec42-kube-api-access-f69hj\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:21.286310 master-0 kubenswrapper[24928]: I1205 10:47:21.286181 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-multus-daemon-config\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.286310 master-0 kubenswrapper[24928]: I1205 10:47:21.286200 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-serving-ca\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.286310 master-0 kubenswrapper[24928]: I1205 10:47:21.286246 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-ssl-certs\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:21.286310 master-0 kubenswrapper[24928]: I1205 10:47:21.286282 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eb290494-a456-4f0e-9afc-f20abab1a1bf-trusted-ca\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:47:21.286310 master-0 kubenswrapper[24928]: I1205 10:47:21.286311 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-config\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:21.286595 master-0 kubenswrapper[24928]: I1205 10:47:21.286344 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-tls\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.286595 master-0 kubenswrapper[24928]: I1205 10:47:21.286374 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11f563d5-89bb-433c-956a-6d5d2492e8f1-serving-cert\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:47:21.286595 master-0 kubenswrapper[24928]: I1205 10:47:21.286405 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-dir\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.286595 master-0 kubenswrapper[24928]: I1205 10:47:21.286475 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88cccb5b-1ad1-4fab-b34c-90252794ee20-serving-cert\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:47:21.286595 master-0 kubenswrapper[24928]: I1205 10:47:21.286552 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4jgc\" (UniqueName: \"kubernetes.io/projected/0a4db531-8af7-4085-a6b2-0de51b527ce6-kube-api-access-g4jgc\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:47:21.286830 master-0 kubenswrapper[24928]: I1205 10:47:21.286616 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/8c649a16-c187-412e-b5da-62a00bee38ab-kube-api-access-d8pv2\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:47:21.286830 master-0 kubenswrapper[24928]: I1205 10:47:21.286643 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-env-overrides\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:47:21.286830 master-0 kubenswrapper[24928]: I1205 10:47:21.286667 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.286830 master-0 kubenswrapper[24928]: I1205 10:47:21.286687 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-cni-binary-copy\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.286830 master-0 kubenswrapper[24928]: I1205 10:47:21.286716 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/11f563d5-89bb-433c-956a-6d5d2492e8f1-serving-cert\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:47:21.286830 master-0 kubenswrapper[24928]: I1205 10:47:21.286769 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:47:21.287148 master-0 kubenswrapper[24928]: I1205 10:47:21.286849 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-env-overrides\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:47:21.287148 master-0 kubenswrapper[24928]: I1205 10:47:21.286890 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-config\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:21.287148 master-0 kubenswrapper[24928]: I1205 10:47:21.286985 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/6a9f011f-36f1-4308-a365-69425c186c7f-cni-binary-copy\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.287148 master-0 kubenswrapper[24928]: I1205 10:47:21.286987 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-sys\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.287148 master-0 kubenswrapper[24928]: I1205 10:47:21.287045 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-metrics-certs\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:47:21.287148 master-0 kubenswrapper[24928]: I1205 10:47:21.287057 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:21.287148 master-0 kubenswrapper[24928]: I1205 10:47:21.287079 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eb290494-a456-4f0e-9afc-f20abab1a1bf-trusted-ca\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:47:21.287148 master-0 kubenswrapper[24928]: I1205 10:47:21.287097 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7824l\" (UniqueName: \"kubernetes.io/projected/da9d48c9-6346-4c9f-a690-f7419499c3e6-kube-api-access-7824l\") pod \"network-check-source-85d8db45d4-c2mhw\" (UID: \"da9d48c9-6346-4c9f-a690-f7419499c3e6\") " pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" Dec 05 10:47:21.287501 master-0 kubenswrapper[24928]: I1205 10:47:21.287179 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-webhook-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:21.287501 master-0 kubenswrapper[24928]: I1205 10:47:21.287211 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8wnp\" (UniqueName: \"kubernetes.io/projected/2c3fab4d-05b8-42ec-8c91-91ff64562649-kube-api-access-l8wnp\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:21.287501 master-0 kubenswrapper[24928]: I1205 10:47:21.287235 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkqz7\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-kube-api-access-vkqz7\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:47:21.287501 master-0 kubenswrapper[24928]: I1205 10:47:21.287259 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-etc-tuned\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.287501 master-0 kubenswrapper[24928]: I1205 10:47:21.287351 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-cnibin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.287501 master-0 kubenswrapper[24928]: I1205 10:47:21.287380 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:21.287501 master-0 kubenswrapper[24928]: I1205 10:47:21.287409 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-catalog-content\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:47:21.287501 master-0 kubenswrapper[24928]: I1205 10:47:21.287455 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwqkb\" (UniqueName: \"kubernetes.io/projected/1a0235af-2cf2-4ad4-b419-764fb56a0107-kube-api-access-cwqkb\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:47:21.287501 master-0 kubenswrapper[24928]: I1205 10:47:21.287471 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-tuned\" (UniqueName: \"kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-etc-tuned\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.287870 master-0 kubenswrapper[24928]: I1205 10:47:21.287575 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-catalog-content\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:47:21.287983 master-0 kubenswrapper[24928]: I1205 10:47:21.287950 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:21.288054 master-0 kubenswrapper[24928]: I1205 10:47:21.288004 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:21.288054 master-0 kubenswrapper[24928]: I1205 10:47:21.288035 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:47:21.288216 master-0 kubenswrapper[24928]: I1205 10:47:21.288115 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-catalog-content\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:47:21.288216 master-0 kubenswrapper[24928]: I1205 10:47:21.288149 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/62f174ff-9f3c-4051-a34a-e55a59207171-machine-approver-tls\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:47:21.288216 master-0 kubenswrapper[24928]: I1205 10:47:21.288174 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/3aa9a063-322b-4dc6-a724-05a66026160b-audit-log\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:21.288451 master-0 kubenswrapper[24928]: I1205 10:47:21.288214 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-catalog-content\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:47:21.288451 master-0 kubenswrapper[24928]: I1205 10:47:21.288238 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-utilities\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:47:21.288451 master-0 kubenswrapper[24928]: I1205 10:47:21.288272 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-catalog-content\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:47:21.288451 master-0 kubenswrapper[24928]: I1205 10:47:21.288295 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/3aa9a063-322b-4dc6-a724-05a66026160b-audit-log\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:21.288451 master-0 kubenswrapper[24928]: I1205 10:47:21.288302 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg9zq\" (UniqueName: \"kubernetes.io/projected/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-kube-api-access-wg9zq\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:47:21.288451 master-0 kubenswrapper[24928]: I1205 10:47:21.288347 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6z8n\" (UniqueName: \"kubernetes.io/projected/4c2975ec-e33d-4960-a708-277d41c79b15-kube-api-access-d6z8n\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:47:21.288451 master-0 kubenswrapper[24928]: I1205 10:47:21.288372 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-env-overrides\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.288451 master-0 kubenswrapper[24928]: I1205 10:47:21.288390 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-catalog-content\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:47:21.288451 master-0 kubenswrapper[24928]: I1205 10:47:21.288400 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cnibin\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.288451 master-0 kubenswrapper[24928]: I1205 10:47:21.288449 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rlts\" (UniqueName: \"kubernetes.io/projected/3aa9a063-322b-4dc6-a724-05a66026160b-kube-api-access-5rlts\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288306 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/489b6771-aa16-469f-9883-9a9233d3c379-utilities\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288553 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1fd0a349-0f66-4c85-95e4-8bbc96648c05-kube-api-access\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288589 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-ca-certs\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288615 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cco-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cco-trusted-ca\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288637 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cert\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288659 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f76d12f-5406-47e2-8337-2f50e35376d6-config\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288682 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-kubernetes\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288728 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4825316a-ea9f-4d3d-838b-fa809a6e49c7-serving-cert\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288789 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-trusted-ca-bundle\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288848 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-env-overrides\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288951 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4825316a-ea9f-4d3d-838b-fa809a6e49c7-serving-cert\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:47:21.289119 master-0 kubenswrapper[24928]: I1205 10:47:21.288959 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f76d12f-5406-47e2-8337-2f50e35376d6-config\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:47:21.289789 master-0 kubenswrapper[24928]: I1205 10:47:21.288828 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-host\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.289789 master-0 kubenswrapper[24928]: I1205 10:47:21.289299 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalogserver-certs\" (UniqueName: \"kubernetes.io/secret/a4fd453c-e667-4bdc-aa9e-3d95ff707200-catalogserver-certs\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.289789 master-0 kubenswrapper[24928]: I1205 10:47:21.289333 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit-dir\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.289789 master-0 kubenswrapper[24928]: I1205 10:47:21.289359 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/21a9bd5a-3ade-48ef-8004-e0492daa85ba-metrics-client-ca\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.289789 master-0 kubenswrapper[24928]: I1205 10:47:21.289386 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f563d5-89bb-433c-956a-6d5d2492e8f1-config\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:47:21.289789 master-0 kubenswrapper[24928]: I1205 10:47:21.289414 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn7rj\" (UniqueName: \"kubernetes.io/projected/eb290494-a456-4f0e-9afc-f20abab1a1bf-kube-api-access-qn7rj\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:47:21.289789 master-0 kubenswrapper[24928]: I1205 10:47:21.289653 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11f563d5-89bb-433c-956a-6d5d2492e8f1-config\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:47:21.289789 master-0 kubenswrapper[24928]: I1205 10:47:21.289701 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-tmp\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.290184 master-0 kubenswrapper[24928]: I1205 10:47:21.289588 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/33616deb-ea10-4a38-8681-ab023b526b11-tmp\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.290184 master-0 kubenswrapper[24928]: I1205 10:47:21.290028 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsrkp\" (UniqueName: \"kubernetes.io/projected/33616deb-ea10-4a38-8681-ab023b526b11-kube-api-access-rsrkp\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.290184 master-0 kubenswrapper[24928]: I1205 10:47:21.290056 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:47:21.290184 master-0 kubenswrapper[24928]: I1205 10:47:21.290088 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-node-log\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.290184 master-0 kubenswrapper[24928]: I1205 10:47:21.290117 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-apiservice-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:21.290184 master-0 kubenswrapper[24928]: I1205 10:47:21.290142 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-serving-cert\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:47:21.290184 master-0 kubenswrapper[24928]: I1205 10:47:21.290168 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-utilities\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290199 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290228 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290261 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290287 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f76d12f-5406-47e2-8337-2f50e35376d6-serving-cert\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290312 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/592373ae-a952-4dd3-a9bc-f9c9c19c0802-cache\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290337 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-utilities\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290377 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g88bl\" (UniqueName: \"kubernetes.io/projected/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-kube-api-access-g88bl\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290409 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ml74w\" (UniqueName: \"kubernetes.io/projected/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f-kube-api-access-ml74w\") pod \"csi-snapshot-controller-6b958b6f94-lgn6v\" (UID: \"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290444 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290464 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-os-release\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290483 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290502 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-784mb\" (UniqueName: \"kubernetes.io/projected/6728a458-098b-44f9-8c8f-b0d76c5825fa-kube-api-access-784mb\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290509 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/592373ae-a952-4dd3-a9bc-f9c9c19c0802-cache\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:21.290537 master-0 kubenswrapper[24928]: I1205 10:47:21.290520 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-serving-cert\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:47:21.290986 master-0 kubenswrapper[24928]: I1205 10:47:21.290554 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-systemd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.290986 master-0 kubenswrapper[24928]: I1205 10:47:21.290647 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7b29f89-e42d-4e53-ad14-05efdce933f0-image-registry-operator-tls\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:47:21.290986 master-0 kubenswrapper[24928]: I1205 10:47:21.290715 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.290986 master-0 kubenswrapper[24928]: I1205 10:47:21.290792 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgtnt\" (UniqueName: \"kubernetes.io/projected/384f1c47-f511-4b58-aa7f-71aef6ef91a9-kube-api-access-lgtnt\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:21.290986 master-0 kubenswrapper[24928]: I1205 10:47:21.290813 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-tuning-operator-tls\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-node-tuning-operator-tls\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:47:21.290986 master-0 kubenswrapper[24928]: I1205 10:47:21.290872 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv4zs\" (UniqueName: \"kubernetes.io/projected/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-kube-api-access-jv4zs\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:47:21.290986 master-0 kubenswrapper[24928]: I1205 10:47:21.290965 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-564nt\" (UniqueName: \"kubernetes.io/projected/baee05cb-62ad-4eda-8a13-e317eae840e5-kube-api-access-564nt\") pod \"node-resolver-qkccw\" (UID: \"baee05cb-62ad-4eda-8a13-e317eae840e5\") " pod="openshift-dns/node-resolver-qkccw" Dec 05 10:47:21.291170 master-0 kubenswrapper[24928]: I1205 10:47:21.290995 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:47:21.291170 master-0 kubenswrapper[24928]: I1205 10:47:21.291019 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wfsv\" (UniqueName: \"kubernetes.io/projected/1e69ce9e-4e6f-4015-9ba6-5a7942570190-kube-api-access-7wfsv\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:47:21.291170 master-0 kubenswrapper[24928]: I1205 10:47:21.291040 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpgqz\" (UniqueName: \"kubernetes.io/projected/d9ff1ae2-8547-42db-9fd0-f4782589ca18-kube-api-access-wpgqz\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:47:21.291170 master-0 kubenswrapper[24928]: I1205 10:47:21.291062 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdc7g\" (UniqueName: \"kubernetes.io/projected/0de26264-46c2-4d29-97da-25a1681d6a8e-kube-api-access-cdc7g\") pod \"cluster-samples-operator-797cfd8b47-glpx7\" (UID: \"0de26264-46c2-4d29-97da-25a1681d6a8e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:47:21.291170 master-0 kubenswrapper[24928]: I1205 10:47:21.291079 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vp9b\" (UniqueName: \"kubernetes.io/projected/49051e6e-5a2f-45c8-bad0-374514a91c07-kube-api-access-6vp9b\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:47:21.291170 master-0 kubenswrapper[24928]: I1205 10:47:21.291153 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-baremetal-operator-tls\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cluster-baremetal-operator-tls\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:21.291334 master-0 kubenswrapper[24928]: I1205 10:47:21.291201 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-ovnkube-config\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:47:21.291334 master-0 kubenswrapper[24928]: I1205 10:47:21.291232 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-trusted-ca\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:47:21.291334 master-0 kubenswrapper[24928]: I1205 10:47:21.291259 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-key\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:47:21.291334 master-0 kubenswrapper[24928]: I1205 10:47:21.291284 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-slash\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.291485 master-0 kubenswrapper[24928]: I1205 10:47:21.291389 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-host-etc-kube\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:21.291485 master-0 kubenswrapper[24928]: I1205 10:47:21.291413 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b760849c-8d83-47da-8677-68445c143bef-ovnkube-config\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:47:21.291485 master-0 kubenswrapper[24928]: I1205 10:47:21.291411 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f76d12f-5406-47e2-8337-2f50e35376d6-serving-cert\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:47:21.291485 master-0 kubenswrapper[24928]: I1205 10:47:21.291455 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-binary-copy\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.291668 master-0 kubenswrapper[24928]: I1205 10:47:21.291500 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-socket-dir-parent\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.291668 master-0 kubenswrapper[24928]: I1205 10:47:21.291525 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-key\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:47:21.291668 master-0 kubenswrapper[24928]: I1205 10:47:21.291556 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-netns\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.291668 master-0 kubenswrapper[24928]: I1205 10:47:21.291587 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-tls\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:47:21.291668 master-0 kubenswrapper[24928]: I1205 10:47:21.291611 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-serving-cert\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:47:21.291668 master-0 kubenswrapper[24928]: I1205 10:47:21.291634 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-encryption-config\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.291668 master-0 kubenswrapper[24928]: I1205 10:47:21.291653 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:47:21.291668 master-0 kubenswrapper[24928]: I1205 10:47:21.291670 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d9ff1ae2-8547-42db-9fd0-f4782589ca18-metrics-client-ca\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:47:21.291962 master-0 kubenswrapper[24928]: I1205 10:47:21.291690 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lwgq\" (UniqueName: \"kubernetes.io/projected/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-kube-api-access-2lwgq\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:47:21.291962 master-0 kubenswrapper[24928]: I1205 10:47:21.291705 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-binary-copy\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.291962 master-0 kubenswrapper[24928]: I1205 10:47:21.291710 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98qt8\" (UniqueName: \"kubernetes.io/projected/c22d947f-a5b6-4f24-b142-dd201c46293b-kube-api-access-98qt8\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:47:21.291962 master-0 kubenswrapper[24928]: I1205 10:47:21.291741 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6728a458-098b-44f9-8c8f-b0d76c5825fa-mcc-auth-proxy-config\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:47:21.291962 master-0 kubenswrapper[24928]: I1205 10:47:21.291761 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-sys\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.291962 master-0 kubenswrapper[24928]: I1205 10:47:21.291791 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-trusted-ca\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:47:21.291962 master-0 kubenswrapper[24928]: I1205 10:47:21.291869 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-serving-cert\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:47:21.292145 master-0 kubenswrapper[24928]: I1205 10:47:21.291978 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-serving-cert\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:21.292145 master-0 kubenswrapper[24928]: I1205 10:47:21.292049 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 10:47:21.292145 master-0 kubenswrapper[24928]: I1205 10:47:21.292081 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a1bdc70-6412-47e0-8330-04d796cc8d55-metrics-tls\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:47:21.292145 master-0 kubenswrapper[24928]: I1205 10:47:21.292057 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:47:21.292145 master-0 kubenswrapper[24928]: I1205 10:47:21.292137 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f132702-304b-46ff-b428-6df1deeffec3-cert\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:47:21.292280 master-0 kubenswrapper[24928]: I1205 10:47:21.292167 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-containers\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:21.292280 master-0 kubenswrapper[24928]: I1205 10:47:21.292202 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-service-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:21.292280 master-0 kubenswrapper[24928]: I1205 10:47:21.292208 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-serving-cert\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:21.292280 master-0 kubenswrapper[24928]: I1205 10:47:21.292240 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-kubelet\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.292389 master-0 kubenswrapper[24928]: I1205 10:47:21.292292 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xjx8\" (UniqueName: \"kubernetes.io/projected/01488ac4-313d-48d5-9e86-7b2011b9e91e-kube-api-access-2xjx8\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:47:21.292389 master-0 kubenswrapper[24928]: I1205 10:47:21.292320 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:21.292389 master-0 kubenswrapper[24928]: I1205 10:47:21.292350 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e69ce9e-4e6f-4015-9ba6-5a7942570190-serving-cert\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:47:21.292389 master-0 kubenswrapper[24928]: I1205 10:47:21.292375 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-client\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:21.292531 master-0 kubenswrapper[24928]: I1205 10:47:21.292401 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-rootfs\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:47:21.292531 master-0 kubenswrapper[24928]: I1205 10:47:21.292446 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvw27\" (UniqueName: \"kubernetes.io/projected/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-kube-api-access-rvw27\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:47:21.292531 master-0 kubenswrapper[24928]: I1205 10:47:21.292503 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a4fd453c-e667-4bdc-aa9e-3d95ff707200-cache\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.292531 master-0 kubenswrapper[24928]: I1205 10:47:21.292528 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-docker\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:21.292672 master-0 kubenswrapper[24928]: I1205 10:47:21.292556 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:21.292672 master-0 kubenswrapper[24928]: I1205 10:47:21.292584 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-etc-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.292672 master-0 kubenswrapper[24928]: I1205 10:47:21.292592 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/a4fd453c-e667-4bdc-aa9e-3d95ff707200-cache\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.292672 master-0 kubenswrapper[24928]: I1205 10:47:21.292609 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-client-ca\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:47:21.292672 master-0 kubenswrapper[24928]: I1205 10:47:21.292633 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operand-assets\" (UniqueName: \"kubernetes.io/empty-dir/49051e6e-5a2f-45c8-bad0-374514a91c07-operand-assets\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:47:21.292672 master-0 kubenswrapper[24928]: I1205 10:47:21.292659 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-bin\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.292829 master-0 kubenswrapper[24928]: I1205 10:47:21.292681 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-config\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:47:21.292829 master-0 kubenswrapper[24928]: I1205 10:47:21.292705 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.292829 master-0 kubenswrapper[24928]: I1205 10:47:21.292736 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5p8s\" (UniqueName: \"kubernetes.io/projected/62f174ff-9f3c-4051-a34a-e55a59207171-kube-api-access-r5p8s\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:47:21.292829 master-0 kubenswrapper[24928]: I1205 10:47:21.292753 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1e69ce9e-4e6f-4015-9ba6-5a7942570190-serving-cert\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:47:21.292829 master-0 kubenswrapper[24928]: I1205 10:47:21.292772 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:21.292829 master-0 kubenswrapper[24928]: I1205 10:47:21.292786 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysconfig\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.292829 master-0 kubenswrapper[24928]: I1205 10:47:21.292812 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-multus-certs\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.293010 master-0 kubenswrapper[24928]: I1205 10:47:21.292841 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/283122ba-be1c-4516-bd0f-df41c13c098b-iptables-alerter-script\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:47:21.293010 master-0 kubenswrapper[24928]: I1205 10:47:21.292884 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58baad85-de54-49e7-a13e-d470d9c50d11-config-volume\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:47:21.293010 master-0 kubenswrapper[24928]: I1205 10:47:21.292983 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv2tg\" (UniqueName: \"kubernetes.io/projected/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-api-access-sv2tg\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:21.293102 master-0 kubenswrapper[24928]: I1205 10:47:21.293015 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/82464e92-4de2-48f0-8772-a489abb16898-machine-api-operator-tls\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:47:21.293102 master-0 kubenswrapper[24928]: I1205 10:47:21.293031 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-config\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:47:21.293102 master-0 kubenswrapper[24928]: I1205 10:47:21.292984 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operand-assets\" (UniqueName: \"kubernetes.io/empty-dir/49051e6e-5a2f-45c8-bad0-374514a91c07-operand-assets\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:47:21.293102 master-0 kubenswrapper[24928]: I1205 10:47:21.293068 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7b29f89-e42d-4e53-ad14-05efdce933f0-trusted-ca\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:47:21.293102 master-0 kubenswrapper[24928]: I1205 10:47:21.293090 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pvrn\" (UniqueName: \"kubernetes.io/projected/489b6771-aa16-469f-9883-9a9233d3c379-kube-api-access-9pvrn\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:47:21.293256 master-0 kubenswrapper[24928]: I1205 10:47:21.293127 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-serving-cert\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.293256 master-0 kubenswrapper[24928]: I1205 10:47:21.293186 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/444f8808-e454-4015-9e20-429e715a08c7-serving-cert\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:47:21.293256 master-0 kubenswrapper[24928]: I1205 10:47:21.293212 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.293256 master-0 kubenswrapper[24928]: I1205 10:47:21.293197 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/283122ba-be1c-4516-bd0f-df41c13c098b-iptables-alerter-script\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:47:21.293367 master-0 kubenswrapper[24928]: I1205 10:47:21.293276 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-client\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:21.293367 master-0 kubenswrapper[24928]: I1205 10:47:21.293331 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-os-release\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.293455 master-0 kubenswrapper[24928]: I1205 10:47:21.293413 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7b29f89-e42d-4e53-ad14-05efdce933f0-trusted-ca\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:47:21.293455 master-0 kubenswrapper[24928]: I1205 10:47:21.293411 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-serving-cert\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:21.293516 master-0 kubenswrapper[24928]: I1205 10:47:21.293499 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/baee05cb-62ad-4eda-8a13-e317eae840e5-hosts-file\") pod \"node-resolver-qkccw\" (UID: \"baee05cb-62ad-4eda-8a13-e317eae840e5\") " pod="openshift-dns/node-resolver-qkccw" Dec 05 10:47:21.293552 master-0 kubenswrapper[24928]: I1205 10:47:21.293514 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.293668 master-0 kubenswrapper[24928]: I1205 10:47:21.293635 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-catalog-content\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:47:21.293709 master-0 kubenswrapper[24928]: I1205 10:47:21.293675 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:47:21.293753 master-0 kubenswrapper[24928]: I1205 10:47:21.293700 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-ca-certs\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.293782 master-0 kubenswrapper[24928]: I1205 10:47:21.293763 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-utilities\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:47:21.293782 master-0 kubenswrapper[24928]: I1205 10:47:21.293770 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/444f8808-e454-4015-9e20-429e715a08c7-serving-cert\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:47:21.293840 master-0 kubenswrapper[24928]: I1205 10:47:21.293785 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn6sc\" (UniqueName: \"kubernetes.io/projected/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-kube-api-access-nn6sc\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:47:21.293840 master-0 kubenswrapper[24928]: I1205 10:47:21.293798 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-catalog-content\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:47:21.293898 master-0 kubenswrapper[24928]: I1205 10:47:21.293837 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-certs\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:47:21.293927 master-0 kubenswrapper[24928]: I1205 10:47:21.293901 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a4db531-8af7-4085-a6b2-0de51b527ce6-utilities\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:47:21.293927 master-0 kubenswrapper[24928]: I1205 10:47:21.293911 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrwt6\" (UniqueName: \"kubernetes.io/projected/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-kube-api-access-lrwt6\") pod \"control-plane-machine-set-operator-7df95c79b5-qnq6t\" (UID: \"24444474-aa9a-4a0f-8b4d-90f0009e0dc7\") " pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:47:21.293985 master-0 kubenswrapper[24928]: I1205 10:47:21.293937 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-service-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:47:21.293985 master-0 kubenswrapper[24928]: I1205 10:47:21.293958 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dk2b\" (UniqueName: \"kubernetes.io/projected/82464e92-4de2-48f0-8772-a489abb16898-kube-api-access-4dk2b\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:47:21.293985 master-0 kubenswrapper[24928]: I1205 10:47:21.293896 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/926263c4-ec5b-41cb-9c30-0c88f636035f-marketplace-operator-metrics\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:47:21.294230 master-0 kubenswrapper[24928]: I1205 10:47:21.294196 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-policies\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.294271 master-0 kubenswrapper[24928]: I1205 10:47:21.294245 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdsh9\" (UniqueName: \"kubernetes.io/projected/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a-kube-api-access-pdsh9\") pod \"migrator-74b7b57c65-sfvzd\" (UID: \"3c4f767d-65f9-434b-8ddd-ceb0b91ab99a\") " pod="openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd" Dec 05 10:47:21.294300 master-0 kubenswrapper[24928]: I1205 10:47:21.294275 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.294328 master-0 kubenswrapper[24928]: I1205 10:47:21.294311 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fbtj\" (UniqueName: \"kubernetes.io/projected/58b95ae4-7f74-4777-8441-0a0ae28199e9-kube-api-access-7fbtj\") pod \"cluster-storage-operator-dcf7fc84b-9rzps\" (UID: \"58b95ae4-7f74-4777-8441-0a0ae28199e9\") " pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:47:21.294328 master-0 kubenswrapper[24928]: I1205 10:47:21.294318 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-service-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:47:21.294388 master-0 kubenswrapper[24928]: I1205 10:47:21.294340 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljdt5\" (UniqueName: \"kubernetes.io/projected/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-kube-api-access-ljdt5\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:21.294388 master-0 kubenswrapper[24928]: I1205 10:47:21.294372 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmhhw\" (UniqueName: \"kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-kube-api-access-kmhhw\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.294472 master-0 kubenswrapper[24928]: I1205 10:47:21.294400 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flvlq\" (UniqueName: \"kubernetes.io/projected/568e9bed-cbd2-49bd-84af-9e17f95a003f-kube-api-access-flvlq\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:47:21.294472 master-0 kubenswrapper[24928]: I1205 10:47:21.294455 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/445d75af-d072-4fa0-91a7-f3fa579b9ca9-service-ca-bundle\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:21.294550 master-0 kubenswrapper[24928]: I1205 10:47:21.294484 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-config\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:47:21.294550 master-0 kubenswrapper[24928]: I1205 10:47:21.294512 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-config\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:47:21.294550 master-0 kubenswrapper[24928]: I1205 10:47:21.294537 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twh85\" (UniqueName: \"kubernetes.io/projected/f7a08359-0379-4364-8b0c-ddb58ff605f4-kube-api-access-twh85\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:21.294747 master-0 kubenswrapper[24928]: I1205 10:47:21.294666 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-root\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.294796 master-0 kubenswrapper[24928]: I1205 10:47:21.294755 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-config\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:47:21.294870 master-0 kubenswrapper[24928]: I1205 10:47:21.294835 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-kubelet\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.294944 master-0 kubenswrapper[24928]: I1205 10:47:21.294916 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6728a458-098b-44f9-8c8f-b0d76c5825fa-proxy-tls\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:47:21.294995 master-0 kubenswrapper[24928]: I1205 10:47:21.294972 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-ovn\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.295046 master-0 kubenswrapper[24928]: I1205 10:47:21.295027 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-serving-cert\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:47:21.295106 master-0 kubenswrapper[24928]: I1205 10:47:21.295077 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-images\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:47:21.295169 master-0 kubenswrapper[24928]: I1205 10:47:21.295139 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkmvj\" (UniqueName: \"kubernetes.io/projected/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-kube-api-access-fkmvj\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:47:21.295258 master-0 kubenswrapper[24928]: I1205 10:47:21.295238 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-serving-cert\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:47:21.295298 master-0 kubenswrapper[24928]: I1205 10:47:21.295248 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.295340 master-0 kubenswrapper[24928]: I1205 10:47:21.295310 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-cvo-updatepayloads\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:21.295400 master-0 kubenswrapper[24928]: I1205 10:47:21.295372 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e69ce9e-4e6f-4015-9ba6-5a7942570190-config\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:47:21.295506 master-0 kubenswrapper[24928]: I1205 10:47:21.295475 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-stats-auth\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:21.295561 master-0 kubenswrapper[24928]: I1205 10:47:21.295540 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-7df95c79b5-qnq6t\" (UID: \"24444474-aa9a-4a0f-8b4d-90f0009e0dc7\") " pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:47:21.295616 master-0 kubenswrapper[24928]: I1205 10:47:21.295590 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e69ce9e-4e6f-4015-9ba6-5a7942570190-config\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:47:21.295616 master-0 kubenswrapper[24928]: I1205 10:47:21.295601 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:21.295767 master-0 kubenswrapper[24928]: I1205 10:47:21.295727 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfz87\" (UniqueName: \"kubernetes.io/projected/b760849c-8d83-47da-8677-68445c143bef-kube-api-access-jfz87\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:47:21.295830 master-0 kubenswrapper[24928]: I1205 10:47:21.295801 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-default-certificate\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:21.295873 master-0 kubenswrapper[24928]: I1205 10:47:21.295841 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-multus\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.295956 master-0 kubenswrapper[24928]: I1205 10:47:21.295926 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:21.296002 master-0 kubenswrapper[24928]: I1205 10:47:21.295979 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloud-credential-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cloud-credential-operator-serving-cert\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:47:21.296038 master-0 kubenswrapper[24928]: I1205 10:47:21.296023 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-client-ca\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:21.296088 master-0 kubenswrapper[24928]: I1205 10:47:21.296063 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:47:21.296133 master-0 kubenswrapper[24928]: I1205 10:47:21.296111 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjrgm\" (UniqueName: \"kubernetes.io/projected/87909f47-f2d7-46f8-a1c8-27336cdcce5d-kube-api-access-rjrgm\") pod \"csi-snapshot-controller-operator-6bc8656fdc-vd94f\" (UID: \"87909f47-f2d7-46f8-a1c8-27336cdcce5d\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:47:21.296179 master-0 kubenswrapper[24928]: I1205 10:47:21.296155 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-client\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.296222 master-0 kubenswrapper[24928]: I1205 10:47:21.296198 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-encryption-config\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.296271 master-0 kubenswrapper[24928]: I1205 10:47:21.296238 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:21.296318 master-0 kubenswrapper[24928]: I1205 10:47:21.296277 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:47:21.296351 master-0 kubenswrapper[24928]: I1205 10:47:21.296328 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-trusted-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:47:21.296351 master-0 kubenswrapper[24928]: I1205 10:47:21.296340 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-srv-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:47:21.296406 master-0 kubenswrapper[24928]: I1205 10:47:21.296362 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:47:21.296406 master-0 kubenswrapper[24928]: I1205 10:47:21.296396 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/01488ac4-313d-48d5-9e86-7b2011b9e91e-proxy-tls\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:47:21.296525 master-0 kubenswrapper[24928]: I1205 10:47:21.296458 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-ovnkube-identity-cm\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:47:21.296525 master-0 kubenswrapper[24928]: I1205 10:47:21.296496 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-bound-sa-token\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:47:21.296589 master-0 kubenswrapper[24928]: I1205 10:47:21.296533 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-config\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.296589 master-0 kubenswrapper[24928]: I1205 10:47:21.296570 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjgc4\" (UniqueName: \"kubernetes.io/projected/495ba1ea-f844-43ec-8be7-47e738f5428a-kube-api-access-tjgc4\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.296724 master-0 kubenswrapper[24928]: I1205 10:47:21.296702 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-ovnkube-identity-cm\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:47:21.296764 master-0 kubenswrapper[24928]: I1205 10:47:21.296724 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/283122ba-be1c-4516-bd0f-df41c13c098b-host-slash\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:47:21.296793 master-0 kubenswrapper[24928]: I1205 10:47:21.296767 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:21.296826 master-0 kubenswrapper[24928]: I1205 10:47:21.296811 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-images\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:47:21.296877 master-0 kubenswrapper[24928]: I1205 10:47:21.296852 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/0babac16-164a-405a-a86c-30524118f046-tls-certificates\") pod \"prometheus-operator-admission-webhook-7c85c4dffd-vjvbz\" (UID: \"0babac16-164a-405a-a86c-30524118f046\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" Dec 05 10:47:21.296911 master-0 kubenswrapper[24928]: I1205 10:47:21.296898 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4825316a-ea9f-4d3d-838b-fa809a6e49c7-config\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:47:21.296944 master-0 kubenswrapper[24928]: I1205 10:47:21.296900 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-trusted-ca-bundle\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:47:21.296972 master-0 kubenswrapper[24928]: I1205 10:47:21.296865 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-config\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.297053 master-0 kubenswrapper[24928]: I1205 10:47:21.297031 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-netd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.297053 master-0 kubenswrapper[24928]: I1205 10:47:21.297046 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4825316a-ea9f-4d3d-838b-fa809a6e49c7-config\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:47:21.297115 master-0 kubenswrapper[24928]: I1205 10:47:21.297067 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-script-lib\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.297115 master-0 kubenswrapper[24928]: I1205 10:47:21.297087 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-containers\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.297115 master-0 kubenswrapper[24928]: I1205 10:47:21.297112 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssssf\" (UniqueName: \"kubernetes.io/projected/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-kube-api-access-ssssf\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:47:21.297203 master-0 kubenswrapper[24928]: I1205 10:47:21.297140 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltjt6\" (UniqueName: \"kubernetes.io/projected/ca45c52e-fb30-4e7c-8c3f-e685c0909916-kube-api-access-ltjt6\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.297203 master-0 kubenswrapper[24928]: I1205 10:47:21.297160 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fd0a349-0f66-4c85-95e4-8bbc96648c05-serving-cert\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:21.297203 master-0 kubenswrapper[24928]: I1205 10:47:21.297180 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-olm-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/49051e6e-5a2f-45c8-bad0-374514a91c07-cluster-olm-operator-serving-cert\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:47:21.297203 master-0 kubenswrapper[24928]: I1205 10:47:21.297200 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4lbg\" (UniqueName: \"kubernetes.io/projected/88cccb5b-1ad1-4fab-b34c-90252794ee20-kube-api-access-f4lbg\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:47:21.297311 master-0 kubenswrapper[24928]: I1205 10:47:21.297218 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/495ba1ea-f844-43ec-8be7-47e738f5428a-ovnkube-script-lib\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.297311 master-0 kubenswrapper[24928]: I1205 10:47:21.297220 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/1a0235af-2cf2-4ad4-b419-764fb56a0107-host-etc-kube\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:47:21.297311 master-0 kubenswrapper[24928]: I1205 10:47:21.297260 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hskm2\" (UniqueName: \"kubernetes.io/projected/445d75af-d072-4fa0-91a7-f3fa579b9ca9-kube-api-access-hskm2\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:21.297311 master-0 kubenswrapper[24928]: I1205 10:47:21.297282 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvnxf\" (UniqueName: \"kubernetes.io/projected/6a9f011f-36f1-4308-a365-69425c186c7f-kube-api-access-mvnxf\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.297439 master-0 kubenswrapper[24928]: I1205 10:47:21.297328 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-profile-collector-cert\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:47:21.297439 master-0 kubenswrapper[24928]: I1205 10:47:21.297352 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-client\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.297439 master-0 kubenswrapper[24928]: I1205 10:47:21.297369 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-image-import-ca\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.297439 master-0 kubenswrapper[24928]: I1205 10:47:21.297391 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:21.297439 master-0 kubenswrapper[24928]: I1205 10:47:21.297411 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0de26264-46c2-4d29-97da-25a1681d6a8e-samples-operator-tls\") pod \"cluster-samples-operator-797cfd8b47-glpx7\" (UID: \"0de26264-46c2-4d29-97da-25a1681d6a8e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:47:21.297644 master-0 kubenswrapper[24928]: I1205 10:47:21.297528 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:47:21.297644 master-0 kubenswrapper[24928]: I1205 10:47:21.297552 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-bound-sa-token\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:47:21.297644 master-0 kubenswrapper[24928]: I1205 10:47:21.297572 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-conf\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.297644 master-0 kubenswrapper[24928]: I1205 10:47:21.297595 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-config\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.297644 master-0 kubenswrapper[24928]: I1205 10:47:21.297614 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:21.297840 master-0 kubenswrapper[24928]: I1205 10:47:21.297650 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fbss\" (UniqueName: \"kubernetes.io/projected/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-kube-api-access-2fbss\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:47:21.297840 master-0 kubenswrapper[24928]: I1205 10:47:21.297680 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:21.297840 master-0 kubenswrapper[24928]: I1205 10:47:21.297706 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f132702-304b-46ff-b428-6df1deeffec3-auth-proxy-config\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:47:21.297840 master-0 kubenswrapper[24928]: I1205 10:47:21.297732 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:47:21.297840 master-0 kubenswrapper[24928]: I1205 10:47:21.297736 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-profile-collector-cert\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:47:21.297840 master-0 kubenswrapper[24928]: I1205 10:47:21.297753 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-webhook-cert\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:47:21.298120 master-0 kubenswrapper[24928]: I1205 10:47:21.297986 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-webhook-cert\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:47:21.298290 master-0 kubenswrapper[24928]: I1205 10:47:21.298254 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/22676fac-b770-4937-9bee-7478bd1babb7-metrics-tls\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:47:21.298330 master-0 kubenswrapper[24928]: I1205 10:47:21.298277 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-utilities\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:47:21.298330 master-0 kubenswrapper[24928]: I1205 10:47:21.298323 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28qvt\" (UniqueName: \"kubernetes.io/projected/baa812e4-b011-41ee-82ab-8f571f0b7e0a-kube-api-access-28qvt\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:47:21.298384 master-0 kubenswrapper[24928]: I1205 10:47:21.298351 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:21.298447 master-0 kubenswrapper[24928]: I1205 10:47:21.298382 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baa812e4-b011-41ee-82ab-8f571f0b7e0a-utilities\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:47:21.298447 master-0 kubenswrapper[24928]: I1205 10:47:21.298388 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-systemd\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.298525 master-0 kubenswrapper[24928]: I1205 10:47:21.298464 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-run\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.298525 master-0 kubenswrapper[24928]: I1205 10:47:21.298519 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.298577 master-0 kubenswrapper[24928]: I1205 10:47:21.298524 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-olm-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/49051e6e-5a2f-45c8-bad0-374514a91c07-cluster-olm-operator-serving-cert\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:47:21.298624 master-0 kubenswrapper[24928]: I1205 10:47:21.298598 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2c3fab4d-05b8-42ec-8c91-91ff64562649-tmpfs\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:21.298833 master-0 kubenswrapper[24928]: I1205 10:47:21.298808 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/2c3fab4d-05b8-42ec-8c91-91ff64562649-tmpfs\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:21.298900 master-0 kubenswrapper[24928]: I1205 10:47:21.298850 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx7jr\" (UniqueName: \"kubernetes.io/projected/58baad85-de54-49e7-a13e-d470d9c50d11-kube-api-access-jx7jr\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:47:21.299010 master-0 kubenswrapper[24928]: I1205 10:47:21.298983 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd87p\" (UniqueName: \"kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-kube-api-access-rd87p\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:21.299067 master-0 kubenswrapper[24928]: I1205 10:47:21.299026 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-bin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.299067 master-0 kubenswrapper[24928]: I1205 10:47:21.299052 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloud-controller-manager-operator-tls\" (UniqueName: \"kubernetes.io/secret/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-cloud-controller-manager-operator-tls\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:21.299152 master-0 kubenswrapper[24928]: I1205 10:47:21.299078 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"whereabouts-configmap\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-whereabouts-configmap\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.299152 master-0 kubenswrapper[24928]: I1205 10:47:21.299098 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-metrics-certs\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:21.299152 master-0 kubenswrapper[24928]: I1205 10:47:21.299123 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:47:21.299152 master-0 kubenswrapper[24928]: I1205 10:47:21.299143 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c22d947f-a5b6-4f24-b142-dd201c46293b-available-featuregates\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:47:21.299286 master-0 kubenswrapper[24928]: I1205 10:47:21.299162 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4825316a-ea9f-4d3d-838b-fa809a6e49c7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:47:21.299286 master-0 kubenswrapper[24928]: I1205 10:47:21.299183 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ktnb\" (UniqueName: \"kubernetes.io/projected/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-kube-api-access-9ktnb\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:47:21.299286 master-0 kubenswrapper[24928]: I1205 10:47:21.299201 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-system-cni-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.299286 master-0 kubenswrapper[24928]: I1205 10:47:21.299221 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-proxy-tls\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:47:21.299286 master-0 kubenswrapper[24928]: I1205 10:47:21.299243 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkzkh\" (UniqueName: \"kubernetes.io/projected/1f132702-304b-46ff-b428-6df1deeffec3-kube-api-access-vkzkh\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:47:21.299286 master-0 kubenswrapper[24928]: I1205 10:47:21.299262 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-metrics-client-ca\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:21.299286 master-0 kubenswrapper[24928]: I1205 10:47:21.299281 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnzgs\" (UniqueName: \"kubernetes.io/projected/0aa6e138-4b1f-4047-8255-a2b14d044588-kube-api-access-wnzgs\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:47:21.299529 master-0 kubenswrapper[24928]: I1205 10:47:21.299302 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-telemetry-config\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:47:21.299529 master-0 kubenswrapper[24928]: I1205 10:47:21.299320 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-node-pullsecrets\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.299529 master-0 kubenswrapper[24928]: I1205 10:47:21.299339 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-textfile\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.299529 master-0 kubenswrapper[24928]: I1205 10:47:21.299358 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/54f9f9d4-8c31-4636-8010-22ee11b9b323-volume-directive-shadow\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:21.299529 master-0 kubenswrapper[24928]: I1205 10:47:21.299381 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:21.299529 master-0 kubenswrapper[24928]: I1205 10:47:21.299410 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c22d947f-a5b6-4f24-b142-dd201c46293b-serving-cert\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:47:21.299529 master-0 kubenswrapper[24928]: I1205 10:47:21.299456 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrdsv\" (UniqueName: \"kubernetes.io/projected/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-kube-api-access-jrdsv\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.299529 master-0 kubenswrapper[24928]: I1205 10:47:21.299485 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gfgm\" (UniqueName: \"kubernetes.io/projected/3332b604-6f0d-4243-bce1-b3ab76e58c78-kube-api-access-6gfgm\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:47:21.299529 master-0 kubenswrapper[24928]: I1205 10:47:21.299522 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-trusted-ca-bundle\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299542 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a0235af-2cf2-4ad4-b419-764fb56a0107-metrics-tls\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299561 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzwfq\" (UniqueName: \"kubernetes.io/projected/6f76d12f-5406-47e2-8337-2f50e35376d6-kube-api-access-fzwfq\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299582 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-system-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299601 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-snapshots\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299619 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-serving-cert\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299637 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-srv-cert\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299654 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-serving-ca\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299673 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2vd4\" (UniqueName: \"kubernetes.io/projected/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-kube-api-access-h2vd4\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299692 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-k8s-cni-cncf-io\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299708 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-netns\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299722 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-var-lib-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299738 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-log-socket\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299756 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-images\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299772 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-wtmp\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.299778 master-0 kubenswrapper[24928]: I1205 10:47:21.299790 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/444f8808-e454-4015-9e20-429e715a08c7-config\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.299813 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11f563d5-89bb-433c-956a-6d5d2492e8f1-kube-api-access\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.299836 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-etc-kubernetes\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.299861 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtsw9\" (UniqueName: \"kubernetes.io/projected/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-kube-api-access-vtsw9\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.299888 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khfxz\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-kube-api-access-khfxz\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.299912 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/495ba1ea-f844-43ec-8be7-47e738f5428a-ovn-node-metrics-cert\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.299938 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-storage-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/58b95ae4-7f74-4777-8441-0a0ae28199e9-cluster-storage-operator-serving-cert\") pod \"cluster-storage-operator-dcf7fc84b-9rzps\" (UID: \"58b95ae4-7f74-4777-8441-0a0ae28199e9\") " pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.299965 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fncz7\" (UniqueName: \"kubernetes.io/projected/926263c4-ec5b-41cb-9c30-0c88f636035f-kube-api-access-fncz7\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.299993 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22676fac-b770-4937-9bee-7478bd1babb7-trusted-ca\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.300017 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-profile-collector-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.300039 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b760849c-8d83-47da-8677-68445c143bef-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.300057 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1fd0a349-0f66-4c85-95e4-8bbc96648c05-service-ca\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.300076 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-docker\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.300095 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhjhk\" (UniqueName: \"kubernetes.io/projected/283122ba-be1c-4516-bd0f-df41c13c098b-kube-api-access-vhjhk\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.300113 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.300133 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.300150 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.300174 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-conf-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.300199 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-service-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:21.300233 master-0 kubenswrapper[24928]: I1205 10:47:21.300226 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-mcd-auth-proxy-config\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300253 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-config\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300277 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-cabundle\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300303 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-auth-proxy-config\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300329 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgvv6\" (UniqueName: \"kubernetes.io/projected/b7508cd3-7421-4ab9-be1f-318db2853596-kube-api-access-kgvv6\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300355 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95jmp\" (UniqueName: \"kubernetes.io/projected/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-kube-api-access-95jmp\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300381 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300408 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-tls\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300450 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-modprobe-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300478 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-var-lib-kubelet\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300506 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-proxy-ca-bundles\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300535 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8bpw\" (UniqueName: \"kubernetes.io/projected/21a9bd5a-3ade-48ef-8004-e0492daa85ba-kube-api-access-r8bpw\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300562 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/444f8808-e454-4015-9e20-429e715a08c7-kube-api-access\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300591 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-config\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300658 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:21.300745 master-0 kubenswrapper[24928]: I1205 10:47:21.300688 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:21.301279 master-0 kubenswrapper[24928]: I1205 10:47:21.301255 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"whereabouts-configmap\" (UniqueName: \"kubernetes.io/configmap/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-whereabouts-configmap\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.301810 master-0 kubenswrapper[24928]: I1205 10:47:21.301645 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:47:21.302951 master-0 kubenswrapper[24928]: I1205 10:47:21.302456 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/495ba1ea-f844-43ec-8be7-47e738f5428a-ovn-node-metrics-cert\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.302951 master-0 kubenswrapper[24928]: I1205 10:47:21.302484 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/eb290494-a456-4f0e-9afc-f20abab1a1bf-apiservice-cert\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:47:21.302951 master-0 kubenswrapper[24928]: I1205 10:47:21.302545 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-profile-collector-cert\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:47:21.302951 master-0 kubenswrapper[24928]: I1205 10:47:21.302551 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/444f8808-e454-4015-9e20-429e715a08c7-config\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:47:21.302951 master-0 kubenswrapper[24928]: I1205 10:47:21.302609 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22676fac-b770-4937-9bee-7478bd1babb7-trusted-ca\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:47:21.302951 master-0 kubenswrapper[24928]: I1205 10:47:21.302684 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7a08359-0379-4364-8b0c-ddb58ff605f4-etcd-service-ca\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:21.302951 master-0 kubenswrapper[24928]: I1205 10:47:21.302720 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-config\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:47:21.302951 master-0 kubenswrapper[24928]: I1205 10:47:21.302829 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"snapshots\" (UniqueName: \"kubernetes.io/empty-dir/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-snapshots\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:21.302951 master-0 kubenswrapper[24928]: I1205 10:47:21.302853 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1a0235af-2cf2-4ad4-b419-764fb56a0107-metrics-tls\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:47:21.302951 master-0 kubenswrapper[24928]: I1205 10:47:21.302871 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c22d947f-a5b6-4f24-b142-dd201c46293b-available-featuregates\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:47:21.302951 master-0 kubenswrapper[24928]: I1205 10:47:21.302903 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-textfile\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.303275 master-0 kubenswrapper[24928]: I1205 10:47:21.303012 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-telemetry-config\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:47:21.303275 master-0 kubenswrapper[24928]: I1205 10:47:21.303181 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-signing-cabundle\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:47:21.303275 master-0 kubenswrapper[24928]: I1205 10:47:21.303192 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/54f9f9d4-8c31-4636-8010-22ee11b9b323-volume-directive-shadow\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:21.303356 master-0 kubenswrapper[24928]: I1205 10:47:21.303272 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c649a16-c187-412e-b5da-62a00bee38ab-package-server-manager-serving-cert\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:47:21.303523 master-0 kubenswrapper[24928]: I1205 10:47:21.303397 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c22d947f-a5b6-4f24-b142-dd201c46293b-serving-cert\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:47:21.303572 master-0 kubenswrapper[24928]: I1205 10:47:21.303543 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b760849c-8d83-47da-8677-68445c143bef-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:47:21.303736 master-0 kubenswrapper[24928]: I1205 10:47:21.303651 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:21.313817 master-0 kubenswrapper[24928]: I1205 10:47:21.313535 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 10:47:21.317524 master-0 kubenswrapper[24928]: I1205 10:47:21.317389 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58baad85-de54-49e7-a13e-d470d9c50d11-metrics-tls\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:47:21.333547 master-0 kubenswrapper[24928]: I1205 10:47:21.333492 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 10:47:21.353645 master-0 kubenswrapper[24928]: I1205 10:47:21.353553 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 10:47:21.373403 master-0 kubenswrapper[24928]: I1205 10:47:21.373302 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 10:47:21.384645 master-0 kubenswrapper[24928]: I1205 10:47:21.384561 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58baad85-de54-49e7-a13e-d470d9c50d11-config-volume\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:47:21.392593 master-0 kubenswrapper[24928]: I1205 10:47:21.392520 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 10:47:21.399171 master-0 kubenswrapper[24928]: I1205 10:47:21.399121 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-client\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.403014 master-0 kubenswrapper[24928]: I1205 10:47:21.402951 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-docker\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.403112 master-0 kubenswrapper[24928]: I1205 10:47:21.403050 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.403112 master-0 kubenswrapper[24928]: I1205 10:47:21.403081 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-conf-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.403112 master-0 kubenswrapper[24928]: I1205 10:47:21.403077 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-docker\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.403262 master-0 kubenswrapper[24928]: I1205 10:47:21.403216 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-sysctl-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.403262 master-0 kubenswrapper[24928]: I1205 10:47:21.403252 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-modprobe-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.403351 master-0 kubenswrapper[24928]: I1205 10:47:21.403266 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-conf-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.403351 master-0 kubenswrapper[24928]: I1205 10:47:21.403284 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-var-lib-kubelet\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.403594 master-0 kubenswrapper[24928]: I1205 10:47:21.403519 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-modprobe-d\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-modprobe-d\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.403648 master-0 kubenswrapper[24928]: I1205 10:47:21.403610 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-var-lib-kubelet\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.403709 master-0 kubenswrapper[24928]: I1205 10:47:21.403686 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-hostroot\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.403781 master-0 kubenswrapper[24928]: I1205 10:47:21.403757 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:21.403827 master-0 kubenswrapper[24928]: I1205 10:47:21.403791 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-lib-modules\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.403864 master-0 kubenswrapper[24928]: I1205 10:47:21.403829 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:21.403894 master-0 kubenswrapper[24928]: I1205 10:47:21.403797 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-hostroot\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.403923 master-0 kubenswrapper[24928]: I1205 10:47:21.403878 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-systemd-units\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.403923 master-0 kubenswrapper[24928]: I1205 10:47:21.403911 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-systemd-units\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.403987 master-0 kubenswrapper[24928]: I1205 10:47:21.403953 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-ssl-certs\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:21.403987 master-0 kubenswrapper[24928]: I1205 10:47:21.403970 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-lib-modules\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.404042 master-0 kubenswrapper[24928]: I1205 10:47:21.404009 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-dir\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.404114 master-0 kubenswrapper[24928]: I1205 10:47:21.404078 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-dir\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.404114 master-0 kubenswrapper[24928]: I1205 10:47:21.404087 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-ssl-certs\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:21.404194 master-0 kubenswrapper[24928]: I1205 10:47:21.404134 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.404194 master-0 kubenswrapper[24928]: I1205 10:47:21.404185 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-sys\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.404323 master-0 kubenswrapper[24928]: I1205 10:47:21.404291 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.404380 master-0 kubenswrapper[24928]: I1205 10:47:21.404359 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-cnibin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.404433 master-0 kubenswrapper[24928]: I1205 10:47:21.404384 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-sys\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.404433 master-0 kubenswrapper[24928]: I1205 10:47:21.404412 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:21.404515 master-0 kubenswrapper[24928]: I1205 10:47:21.404490 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-cnibin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.404515 master-0 kubenswrapper[24928]: I1205 10:47:21.404500 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cnibin\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.404602 master-0 kubenswrapper[24928]: I1205 10:47:21.404592 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-cnibin\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.404680 master-0 kubenswrapper[24928]: I1205 10:47:21.404632 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:21.404759 master-0 kubenswrapper[24928]: I1205 10:47:21.404730 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-kubernetes\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.404807 master-0 kubenswrapper[24928]: I1205 10:47:21.404788 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-host\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.404875 master-0 kubenswrapper[24928]: I1205 10:47:21.404850 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit-dir\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.404916 master-0 kubenswrapper[24928]: I1205 10:47:21.404879 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-kubernetes\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.405047 master-0 kubenswrapper[24928]: I1205 10:47:21.404954 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-node-log\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.405047 master-0 kubenswrapper[24928]: I1205 10:47:21.404990 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-host\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.405047 master-0 kubenswrapper[24928]: I1205 10:47:21.404955 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit-dir\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.405156 master-0 kubenswrapper[24928]: I1205 10:47:21.405044 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.405156 master-0 kubenswrapper[24928]: I1205 10:47:21.405096 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.405156 master-0 kubenswrapper[24928]: I1205 10:47:21.405043 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-node-log\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.405156 master-0 kubenswrapper[24928]: I1205 10:47:21.405129 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-os-release\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.405303 master-0 kubenswrapper[24928]: I1205 10:47:21.405172 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-systemd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.405303 master-0 kubenswrapper[24928]: I1205 10:47:21.405205 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.405303 master-0 kubenswrapper[24928]: I1205 10:47:21.405254 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-systemd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.405303 master-0 kubenswrapper[24928]: I1205 10:47:21.405272 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-os-release\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.405303 master-0 kubenswrapper[24928]: I1205 10:47:21.405298 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.405552 master-0 kubenswrapper[24928]: I1205 10:47:21.405391 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-slash\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.405552 master-0 kubenswrapper[24928]: I1205 10:47:21.405452 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-host-etc-kube\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:21.405552 master-0 kubenswrapper[24928]: I1205 10:47:21.405484 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-slash\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.405552 master-0 kubenswrapper[24928]: I1205 10:47:21.405488 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-socket-dir-parent\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.405552 master-0 kubenswrapper[24928]: I1205 10:47:21.405534 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-multus-socket-dir-parent\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.405740 master-0 kubenswrapper[24928]: I1205 10:47:21.405569 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-netns\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.405740 master-0 kubenswrapper[24928]: I1205 10:47:21.405591 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-host-etc-kube\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:21.405740 master-0 kubenswrapper[24928]: I1205 10:47:21.405655 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-sys\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.405740 master-0 kubenswrapper[24928]: I1205 10:47:21.405683 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-netns\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.405740 master-0 kubenswrapper[24928]: I1205 10:47:21.405700 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-containers\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:21.405929 master-0 kubenswrapper[24928]: I1205 10:47:21.405768 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-containers\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:21.405929 master-0 kubenswrapper[24928]: I1205 10:47:21.405771 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-sys\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.405929 master-0 kubenswrapper[24928]: I1205 10:47:21.405810 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-kubelet\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.405929 master-0 kubenswrapper[24928]: I1205 10:47:21.405849 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:21.405929 master-0 kubenswrapper[24928]: I1205 10:47:21.405875 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-rootfs\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:47:21.406107 master-0 kubenswrapper[24928]: I1205 10:47:21.405931 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-kubelet\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.406107 master-0 kubenswrapper[24928]: I1205 10:47:21.405990 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-docker\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:21.406107 master-0 kubenswrapper[24928]: I1205 10:47:21.406008 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-rootfs\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:47:21.406107 master-0 kubenswrapper[24928]: I1205 10:47:21.406041 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-etc-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.406107 master-0 kubenswrapper[24928]: I1205 10:47:21.406078 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-docker\" (UniqueName: \"kubernetes.io/host-path/592373ae-a952-4dd3-a9bc-f9c9c19c0802-etc-docker\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:21.406107 master-0 kubenswrapper[24928]: I1205 10:47:21.406099 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-bin\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.406324 master-0 kubenswrapper[24928]: I1205 10:47:21.406139 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-etc-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.406324 master-0 kubenswrapper[24928]: I1205 10:47:21.406159 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-bin\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.406324 master-0 kubenswrapper[24928]: I1205 10:47:21.406193 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysconfig\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.406324 master-0 kubenswrapper[24928]: I1205 10:47:21.406228 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-sysconfig\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysconfig\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.406324 master-0 kubenswrapper[24928]: I1205 10:47:21.406250 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-multus-certs\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.406527 master-0 kubenswrapper[24928]: I1205 10:47:21.406347 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-multus-certs\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.406527 master-0 kubenswrapper[24928]: I1205 10:47:21.406350 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-os-release\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.406527 master-0 kubenswrapper[24928]: I1205 10:47:21.406402 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/baee05cb-62ad-4eda-8a13-e317eae840e5-hosts-file\") pod \"node-resolver-qkccw\" (UID: \"baee05cb-62ad-4eda-8a13-e317eae840e5\") " pod="openshift-dns/node-resolver-qkccw" Dec 05 10:47:21.406527 master-0 kubenswrapper[24928]: I1205 10:47:21.406414 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-os-release\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.406527 master-0 kubenswrapper[24928]: I1205 10:47:21.406514 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.406709 master-0 kubenswrapper[24928]: I1205 10:47:21.406591 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/baee05cb-62ad-4eda-8a13-e317eae840e5-hosts-file\") pod \"node-resolver-qkccw\" (UID: \"baee05cb-62ad-4eda-8a13-e317eae840e5\") " pod="openshift-dns/node-resolver-qkccw" Dec 05 10:47:21.406709 master-0 kubenswrapper[24928]: I1205 10:47:21.406609 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-root\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.406709 master-0 kubenswrapper[24928]: I1205 10:47:21.406633 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.406709 master-0 kubenswrapper[24928]: I1205 10:47:21.406640 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-kubelet\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.406709 master-0 kubenswrapper[24928]: I1205 10:47:21.406666 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-root\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.406709 master-0 kubenswrapper[24928]: I1205 10:47:21.406675 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-kubelet\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.406709 master-0 kubenswrapper[24928]: I1205 10:47:21.406698 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-ovn\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.406998 master-0 kubenswrapper[24928]: I1205 10:47:21.406722 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-run-ovn\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.406998 master-0 kubenswrapper[24928]: I1205 10:47:21.406736 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.406998 master-0 kubenswrapper[24928]: I1205 10:47:21.406756 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-cvo-updatepayloads\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:21.406998 master-0 kubenswrapper[24928]: I1205 10:47:21.406793 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-tuning-conf-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.406998 master-0 kubenswrapper[24928]: I1205 10:47:21.406845 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-multus\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.406998 master-0 kubenswrapper[24928]: I1205 10:47:21.406877 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-multus\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.406998 master-0 kubenswrapper[24928]: I1205 10:47:21.406897 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1fd0a349-0f66-4c85-95e4-8bbc96648c05-etc-cvo-updatepayloads\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:21.406998 master-0 kubenswrapper[24928]: I1205 10:47:21.406962 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/283122ba-be1c-4516-bd0f-df41c13c098b-host-slash\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:47:21.406998 master-0 kubenswrapper[24928]: I1205 10:47:21.406988 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:21.406998 master-0 kubenswrapper[24928]: I1205 10:47:21.406992 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/283122ba-be1c-4516-bd0f-df41c13c098b-host-slash\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:47:21.407372 master-0 kubenswrapper[24928]: I1205 10:47:21.407029 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:21.407372 master-0 kubenswrapper[24928]: I1205 10:47:21.407074 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-netd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.407372 master-0 kubenswrapper[24928]: I1205 10:47:21.407101 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-containers\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.407372 master-0 kubenswrapper[24928]: I1205 10:47:21.407140 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-cni-netd\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.407372 master-0 kubenswrapper[24928]: I1205 10:47:21.407221 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-containers\" (UniqueName: \"kubernetes.io/host-path/a4fd453c-e667-4bdc-aa9e-3d95ff707200-etc-containers\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.407372 master-0 kubenswrapper[24928]: I1205 10:47:21.407284 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/1a0235af-2cf2-4ad4-b419-764fb56a0107-host-etc-kube\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:47:21.407632 master-0 kubenswrapper[24928]: I1205 10:47:21.407395 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/1a0235af-2cf2-4ad4-b419-764fb56a0107-host-etc-kube\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:47:21.407632 master-0 kubenswrapper[24928]: I1205 10:47:21.407539 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-conf\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.407728 master-0 kubenswrapper[24928]: I1205 10:47:21.407645 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:21.407774 master-0 kubenswrapper[24928]: I1205 10:47:21.407739 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:21.407774 master-0 kubenswrapper[24928]: I1205 10:47:21.407744 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-sysctl-conf\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-sysctl-conf\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.407855 master-0 kubenswrapper[24928]: I1205 10:47:21.407784 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-systemd\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.407855 master-0 kubenswrapper[24928]: I1205 10:47:21.407832 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-run\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.407855 master-0 kubenswrapper[24928]: I1205 10:47:21.407867 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-systemd\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-etc-systemd\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.408021 master-0 kubenswrapper[24928]: I1205 10:47:21.407946 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-bin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.408021 master-0 kubenswrapper[24928]: I1205 10:47:21.408003 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-var-lib-cni-bin\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.408098 master-0 kubenswrapper[24928]: I1205 10:47:21.408035 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:21.408098 master-0 kubenswrapper[24928]: I1205 10:47:21.408009 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/33616deb-ea10-4a38-8681-ab023b526b11-run\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:21.408182 master-0 kubenswrapper[24928]: I1205 10:47:21.408101 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-system-cni-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.408255 master-0 kubenswrapper[24928]: I1205 10:47:21.408222 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-system-cni-dir\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:21.408255 master-0 kubenswrapper[24928]: I1205 10:47:21.408245 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-node-pullsecrets\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.408363 master-0 kubenswrapper[24928]: I1205 10:47:21.408343 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/d5198438-06ae-4e63-a7e3-950ba23bba9c-node-pullsecrets\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.408500 master-0 kubenswrapper[24928]: I1205 10:47:21.408474 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-system-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.408604 master-0 kubenswrapper[24928]: I1205 10:47:21.408581 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-k8s-cni-cncf-io\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.408654 master-0 kubenswrapper[24928]: I1205 10:47:21.408613 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-netns\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.408654 master-0 kubenswrapper[24928]: I1205 10:47:21.408663 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-system-cni-dir\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.408766 master-0 kubenswrapper[24928]: I1205 10:47:21.408666 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-host-run-netns\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.408766 master-0 kubenswrapper[24928]: I1205 10:47:21.408680 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-host-run-k8s-cni-cncf-io\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.408766 master-0 kubenswrapper[24928]: I1205 10:47:21.408712 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-var-lib-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.408766 master-0 kubenswrapper[24928]: I1205 10:47:21.408684 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-var-lib-openvswitch\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.408766 master-0 kubenswrapper[24928]: I1205 10:47:21.408754 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-log-socket\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.408938 master-0 kubenswrapper[24928]: I1205 10:47:21.408794 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-wtmp\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.408938 master-0 kubenswrapper[24928]: I1205 10:47:21.408836 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-etc-kubernetes\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.408938 master-0 kubenswrapper[24928]: I1205 10:47:21.408863 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/495ba1ea-f844-43ec-8be7-47e738f5428a-log-socket\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:21.409021 master-0 kubenswrapper[24928]: I1205 10:47:21.408988 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6a9f011f-36f1-4308-a365-69425c186c7f-etc-kubernetes\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:21.409317 master-0 kubenswrapper[24928]: I1205 10:47:21.409274 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-wtmp\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:21.412988 master-0 kubenswrapper[24928]: I1205 10:47:21.412941 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 10:47:21.419910 master-0 kubenswrapper[24928]: I1205 10:47:21.419852 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-serving-cert\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.433746 master-0 kubenswrapper[24928]: I1205 10:47:21.433683 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 10:47:21.435737 master-0 kubenswrapper[24928]: I1205 10:47:21.435679 24928 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 10:47:21.439387 master-0 kubenswrapper[24928]: I1205 10:47:21.439309 24928 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientMemory" Dec 05 10:47:21.439387 master-0 kubenswrapper[24928]: I1205 10:47:21.439382 24928 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasNoDiskPressure" Dec 05 10:47:21.439387 master-0 kubenswrapper[24928]: I1205 10:47:21.439394 24928 kubelet_node_status.go:724] "Recording event message for node" node="master-0" event="NodeHasSufficientPID" Dec 05 10:47:21.439956 master-0 kubenswrapper[24928]: I1205 10:47:21.439899 24928 kubelet_node_status.go:76] "Attempting to register node" node="master-0" Dec 05 10:47:21.442165 master-0 kubenswrapper[24928]: I1205 10:47:21.442097 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-metrics-certs\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:21.452871 master-0 kubenswrapper[24928]: I1205 10:47:21.452776 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 10:47:21.456543 master-0 kubenswrapper[24928]: I1205 10:47:21.456487 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-default-certificate\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:21.465277 master-0 kubenswrapper[24928]: I1205 10:47:21.465196 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:21.465277 master-0 kubenswrapper[24928]: I1205 10:47:21.465252 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:21.472864 master-0 kubenswrapper[24928]: I1205 10:47:21.472587 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 10:47:21.476305 master-0 kubenswrapper[24928]: I1205 10:47:21.476245 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:21.476923 master-0 kubenswrapper[24928]: I1205 10:47:21.476850 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/445d75af-d072-4fa0-91a7-f3fa579b9ca9-stats-auth\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:21.486241 master-0 kubenswrapper[24928]: I1205 10:47:21.485943 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-2-master-0_68058d25-5888-495f-8571-453f0bf919d9/installer/0.log" Dec 05 10:47:21.486241 master-0 kubenswrapper[24928]: I1205 10:47:21.486060 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:21.492863 master-0 kubenswrapper[24928]: I1205 10:47:21.492796 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 10:47:21.502085 master-0 kubenswrapper[24928]: I1205 10:47:21.502024 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d5198438-06ae-4e63-a7e3-950ba23bba9c-encryption-config\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.512740 master-0 kubenswrapper[24928]: I1205 10:47:21.512698 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 10:47:21.519219 master-0 kubenswrapper[24928]: I1205 10:47:21.518924 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-config\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.532723 master-0 kubenswrapper[24928]: I1205 10:47:21.532661 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 10:47:21.543397 master-0 kubenswrapper[24928]: I1205 10:47:21.543311 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-etcd-serving-ca\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.559787 master-0 kubenswrapper[24928]: I1205 10:47:21.559578 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 10:47:21.569792 master-0 kubenswrapper[24928]: I1205 10:47:21.569700 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-trusted-ca-bundle\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.573699 master-0 kubenswrapper[24928]: I1205 10:47:21.573205 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 10:47:21.593994 master-0 kubenswrapper[24928]: I1205 10:47:21.593852 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 10:47:21.613033 master-0 kubenswrapper[24928]: I1205 10:47:21.612874 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir\") pod \"68058d25-5888-495f-8571-453f0bf919d9\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " Dec 05 10:47:21.613033 master-0 kubenswrapper[24928]: I1205 10:47:21.612886 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 10:47:21.613033 master-0 kubenswrapper[24928]: I1205 10:47:21.613020 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock\") pod \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " Dec 05 10:47:21.614673 master-0 kubenswrapper[24928]: I1205 10:47:21.612972 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "68058d25-5888-495f-8571-453f0bf919d9" (UID: "68058d25-5888-495f-8571-453f0bf919d9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:21.614673 master-0 kubenswrapper[24928]: I1205 10:47:21.613086 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock\") pod \"68058d25-5888-495f-8571-453f0bf919d9\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " Dec 05 10:47:21.614673 master-0 kubenswrapper[24928]: I1205 10:47:21.613119 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir\") pod \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " Dec 05 10:47:21.614673 master-0 kubenswrapper[24928]: I1205 10:47:21.613141 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock" (OuterVolumeSpecName: "var-lock") pod "75ba20a1-b322-4ce8-88d2-d8ee210cc27a" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:21.614673 master-0 kubenswrapper[24928]: I1205 10:47:21.613182 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock" (OuterVolumeSpecName: "var-lock") pod "68058d25-5888-495f-8571-453f0bf919d9" (UID: "68058d25-5888-495f-8571-453f0bf919d9"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:21.614673 master-0 kubenswrapper[24928]: I1205 10:47:21.613290 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "75ba20a1-b322-4ce8-88d2-d8ee210cc27a" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:21.614908 master-0 kubenswrapper[24928]: I1205 10:47:21.614868 24928 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:21.614908 master-0 kubenswrapper[24928]: I1205 10:47:21.614900 24928 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:21.615023 master-0 kubenswrapper[24928]: I1205 10:47:21.614923 24928 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/68058d25-5888-495f-8571-453f0bf919d9-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:21.615023 master-0 kubenswrapper[24928]: I1205 10:47:21.614972 24928 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:21.615107 master-0 kubenswrapper[24928]: I1205 10:47:21.615015 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/445d75af-d072-4fa0-91a7-f3fa579b9ca9-service-ca-bundle\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:21.633265 master-0 kubenswrapper[24928]: I1205 10:47:21.633170 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 10:47:21.653017 master-0 kubenswrapper[24928]: I1205 10:47:21.652935 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 10:47:21.658593 master-0 kubenswrapper[24928]: I1205 10:47:21.658524 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-image-import-ca\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.672343 master-0 kubenswrapper[24928]: I1205 10:47:21.672274 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 10:47:21.693544 master-0 kubenswrapper[24928]: I1205 10:47:21.693163 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 10:47:21.695327 master-0 kubenswrapper[24928]: I1205 10:47:21.695233 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d5198438-06ae-4e63-a7e3-950ba23bba9c-audit\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:21.712959 master-0 kubenswrapper[24928]: I1205 10:47:21.712901 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Dec 05 10:47:21.717305 master-0 kubenswrapper[24928]: I1205 10:47:21.717257 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/0babac16-164a-405a-a86c-30524118f046-tls-certificates\") pod \"prometheus-operator-admission-webhook-7c85c4dffd-vjvbz\" (UID: \"0babac16-164a-405a-a86c-30524118f046\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" Dec 05 10:47:21.733562 master-0 kubenswrapper[24928]: I1205 10:47:21.733357 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 10:47:21.734052 master-0 kubenswrapper[24928]: I1205 10:47:21.734012 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-serving-cert\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.753172 master-0 kubenswrapper[24928]: I1205 10:47:21.753029 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 10:47:21.756784 master-0 kubenswrapper[24928]: I1205 10:47:21.756733 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-client\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.772737 master-0 kubenswrapper[24928]: I1205 10:47:21.772654 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 10:47:21.793771 master-0 kubenswrapper[24928]: I1205 10:47:21.793696 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 10:47:21.804152 master-0 kubenswrapper[24928]: I1205 10:47:21.804088 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-trusted-ca-bundle\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.813045 master-0 kubenswrapper[24928]: I1205 10:47:21.812989 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 10:47:21.831949 master-0 kubenswrapper[24928]: I1205 10:47:21.831879 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 10:47:21.836850 master-0 kubenswrapper[24928]: I1205 10:47:21.836653 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ca45c52e-fb30-4e7c-8c3f-e685c0909916-encryption-config\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.852594 master-0 kubenswrapper[24928]: I1205 10:47:21.852544 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 10:47:21.855254 master-0 kubenswrapper[24928]: I1205 10:47:21.855193 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-audit-policies\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.872552 master-0 kubenswrapper[24928]: I1205 10:47:21.872479 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 10:47:21.876790 master-0 kubenswrapper[24928]: I1205 10:47:21.876733 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ca45c52e-fb30-4e7c-8c3f-e685c0909916-etcd-serving-ca\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:21.900459 master-0 kubenswrapper[24928]: I1205 10:47:21.900382 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-catalogd"/"catalogd-trusted-ca-bundle" Dec 05 10:47:21.912382 master-0 kubenswrapper[24928]: I1205 10:47:21.912256 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-catalogd"/"openshift-service-ca.crt" Dec 05 10:47:21.914761 master-0 kubenswrapper[24928]: I1205 10:47:21.914662 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-ca-certs\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.933493 master-0 kubenswrapper[24928]: I1205 10:47:21.933383 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-catalogd"/"catalogserver-cert" Dec 05 10:47:21.940578 master-0 kubenswrapper[24928]: I1205 10:47:21.940531 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalogserver-certs\" (UniqueName: \"kubernetes.io/secret/a4fd453c-e667-4bdc-aa9e-3d95ff707200-catalogserver-certs\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:21.951972 master-0 kubenswrapper[24928]: I1205 10:47:21.951928 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-catalogd"/"kube-root-ca.crt" Dec 05 10:47:21.973330 master-0 kubenswrapper[24928]: I1205 10:47:21.973256 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 10:47:21.979586 master-0 kubenswrapper[24928]: I1205 10:47:21.979517 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fd0a349-0f66-4c85-95e4-8bbc96648c05-serving-cert\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:21.992641 master-0 kubenswrapper[24928]: I1205 10:47:21.992559 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-controller"/"kube-root-ca.crt" Dec 05 10:47:22.012244 master-0 kubenswrapper[24928]: I1205 10:47:22.012128 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-controller"/"openshift-service-ca.crt" Dec 05 10:47:22.040371 master-0 kubenswrapper[24928]: I1205 10:47:22.040304 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-controller"/"operator-controller-trusted-ca-bundle" Dec 05 10:47:22.049512 master-0 kubenswrapper[24928]: I1205 10:47:22.049467 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-ca-certs\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:22.052761 master-0 kubenswrapper[24928]: I1205 10:47:22.052733 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 10:47:22.059116 master-0 kubenswrapper[24928]: I1205 10:47:22.054671 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1fd0a349-0f66-4c85-95e4-8bbc96648c05-service-ca\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:22.074320 master-0 kubenswrapper[24928]: I1205 10:47:22.074259 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 10:47:22.094179 master-0 kubenswrapper[24928]: I1205 10:47:22.094095 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 10:47:22.104211 master-0 kubenswrapper[24928]: I1205 10:47:22.104118 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-serving-cert\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:22.113843 master-0 kubenswrapper[24928]: I1205 10:47:22.113771 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 10:47:22.132973 master-0 kubenswrapper[24928]: I1205 10:47:22.132893 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 10:47:22.138016 master-0 kubenswrapper[24928]: I1205 10:47:22.137961 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-config\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:22.152948 master-0 kubenswrapper[24928]: I1205 10:47:22.152868 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 10:47:22.157175 master-0 kubenswrapper[24928]: I1205 10:47:22.157113 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-client-ca\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:22.172242 master-0 kubenswrapper[24928]: I1205 10:47:22.172170 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 10:47:22.191153 master-0 kubenswrapper[24928]: I1205 10:47:22.191073 24928 request.go:700] Waited for 1.005309408s due to client-side throttling, not priority and fairness, request: GET:https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-controller-manager/configmaps?fieldSelector=metadata.name%3Dopenshift-global-ca&limit=500&resourceVersion=0 Dec 05 10:47:22.202443 master-0 kubenswrapper[24928]: I1205 10:47:22.202346 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 10:47:22.203935 master-0 kubenswrapper[24928]: I1205 10:47:22.203862 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-proxy-ca-bundles\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:22.213862 master-0 kubenswrapper[24928]: I1205 10:47:22.213540 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 10:47:22.217929 master-0 kubenswrapper[24928]: I1205 10:47:22.217880 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88cccb5b-1ad1-4fab-b34c-90252794ee20-serving-cert\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:47:22.233251 master-0 kubenswrapper[24928]: I1205 10:47:22.233192 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 10:47:22.233580 master-0 kubenswrapper[24928]: I1205 10:47:22.233528 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-client-ca\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:47:22.252945 master-0 kubenswrapper[24928]: I1205 10:47:22.252895 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 10:47:22.264186 master-0 kubenswrapper[24928]: I1205 10:47:22.263983 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-config\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:47:22.273280 master-0 kubenswrapper[24928]: I1205 10:47:22.273204 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 10:47:22.285846 master-0 kubenswrapper[24928]: E1205 10:47:22.285782 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/metrics-client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286280 master-0 kubenswrapper[24928]: E1205 10:47:22.285863 24928 configmap.go:193] Couldn't get configMap openshift-cluster-machine-approver/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286280 master-0 kubenswrapper[24928]: E1205 10:47:22.285912 24928 configmap.go:193] Couldn't get configMap openshift-machine-api/cluster-baremetal-operator-images: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286280 master-0 kubenswrapper[24928]: E1205 10:47:22.285907 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-metrics-client-ca podName:b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.785884553 +0000 UTC m=+2.789078424 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-client-ca" (UniqueName: "kubernetes.io/configmap/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-metrics-client-ca") pod "prometheus-operator-6c74d9cb9f-r787z" (UID: "b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286280 master-0 kubenswrapper[24928]: E1205 10:47:22.286061 24928 configmap.go:193] Couldn't get configMap openshift-cloud-controller-manager-operator/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286280 master-0 kubenswrapper[24928]: E1205 10:47:22.286154 24928 configmap.go:193] Couldn't get configMap openshift-insights/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286280 master-0 kubenswrapper[24928]: E1205 10:47:22.286104 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/metrics-client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286655 master-0 kubenswrapper[24928]: E1205 10:47:22.286117 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-auth-proxy-config podName:62f174ff-9f3c-4051-a34a-e55a59207171 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.786039237 +0000 UTC m=+2.789233098 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "auth-proxy-config" (UniqueName: "kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-auth-proxy-config") pod "machine-approver-74d9cbffbc-9jbnk" (UID: "62f174ff-9f3c-4051-a34a-e55a59207171") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286655 master-0 kubenswrapper[24928]: E1205 10:47:22.286345 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-images podName:48bd1d86-a6f2-439f-ab04-6a9a442bec42 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.786320085 +0000 UTC m=+2.789513926 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-images") pod "cluster-baremetal-operator-78f758c7b9-6t2gm" (UID: "48bd1d86-a6f2-439f-ab04-6a9a442bec42") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286655 master-0 kubenswrapper[24928]: E1205 10:47:22.286362 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-trusted-ca-bundle podName:65e06ca6-1f15-43ec-b8dc-a300e1cf83bd nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.786354005 +0000 UTC m=+2.789547856 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-trusted-ca-bundle") pod "insights-operator-55965856b6-2sxv7" (UID: "65e06ca6-1f15-43ec-b8dc-a300e1cf83bd") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286655 master-0 kubenswrapper[24928]: E1205 10:47:22.286390 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-auth-proxy-config podName:af7078c2-ad4f-415b-ba2f-77f5145c3b3f nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.786379296 +0000 UTC m=+2.789573257 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "auth-proxy-config" (UniqueName: "kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-auth-proxy-config") pod "cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" (UID: "af7078c2-ad4f-415b-ba2f-77f5145c3b3f") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286655 master-0 kubenswrapper[24928]: E1205 10:47:22.286406 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-metrics-client-ca podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.786398326 +0000 UTC m=+2.789592177 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-client-ca" (UniqueName: "kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-metrics-client-ca") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286655 master-0 kubenswrapper[24928]: E1205 10:47:22.286491 24928 secret.go:189] Couldn't get secret openshift-monitoring/node-exporter-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.286655 master-0 kubenswrapper[24928]: E1205 10:47:22.286543 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-tls podName:21a9bd5a-3ade-48ef-8004-e0492daa85ba nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.78653595 +0000 UTC m=+2.789729801 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-tls" (UniqueName: "kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-tls") pod "node-exporter-bmqsb" (UID: "21a9bd5a-3ade-48ef-8004-e0492daa85ba") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.286655 master-0 kubenswrapper[24928]: E1205 10:47:22.286598 24928 configmap.go:193] Couldn't get configMap openshift-machine-api/baremetal-kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.286655 master-0 kubenswrapper[24928]: E1205 10:47:22.286630 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-config podName:48bd1d86-a6f2-439f-ab04-6a9a442bec42 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.786622162 +0000 UTC m=+2.789816123 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-config") pod "cluster-baremetal-operator-78f758c7b9-6t2gm" (UID: "48bd1d86-a6f2-439f-ab04-6a9a442bec42") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.287041 master-0 kubenswrapper[24928]: E1205 10:47:22.286990 24928 secret.go:189] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.287224 master-0 kubenswrapper[24928]: E1205 10:47:22.287208 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-node-bootstrap-token podName:568e9bed-cbd2-49bd-84af-9e17f95a003f nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.787176255 +0000 UTC m=+2.790370106 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-node-bootstrap-token") pod "machine-config-server-5t4nn" (UID: "568e9bed-cbd2-49bd-84af-9e17f95a003f") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.287449 master-0 kubenswrapper[24928]: E1205 10:47:22.287360 24928 secret.go:189] Couldn't get secret openshift-monitoring/metrics-client-certs: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.287613 master-0 kubenswrapper[24928]: E1205 10:47:22.287578 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs podName:3aa9a063-322b-4dc6-a724-05a66026160b nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.787537803 +0000 UTC m=+2.790731694 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-metrics-client-certs" (UniqueName: "kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs") pod "metrics-server-7c46d76dff-z8d8z" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.287719 master-0 kubenswrapper[24928]: E1205 10:47:22.287702 24928 secret.go:189] Couldn't get secret openshift-monitoring/telemeter-client-kube-rbac-proxy-config: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.287801 master-0 kubenswrapper[24928]: E1205 10:47:22.287749 24928 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.287879 master-0 kubenswrapper[24928]: E1205 10:47:22.287864 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.787854242 +0000 UTC m=+2.791048093 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-telemeter-client-kube-rbac-proxy-config" (UniqueName: "kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.288041 master-0 kubenswrapper[24928]: E1205 10:47:22.287991 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-webhook-cert podName:2c3fab4d-05b8-42ec-8c91-91ff64562649 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.787955064 +0000 UTC m=+2.791148995 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-webhook-cert") pod "packageserver-d7b67d8cf-krp6c" (UID: "2c3fab4d-05b8-42ec-8c91-91ff64562649") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.288173 master-0 kubenswrapper[24928]: E1205 10:47:22.288137 24928 secret.go:189] Couldn't get secret openshift-monitoring/metrics-server-5ll0c5ruaqfm2: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.288215 master-0 kubenswrapper[24928]: E1205 10:47:22.288160 24928 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.288247 master-0 kubenswrapper[24928]: E1205 10:47:22.288229 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle podName:3aa9a063-322b-4dc6-a724-05a66026160b nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.78820649 +0000 UTC m=+2.791400511 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca-bundle" (UniqueName: "kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle") pod "metrics-server-7c46d76dff-z8d8z" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.288281 master-0 kubenswrapper[24928]: E1205 10:47:22.288270 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert podName:b7508cd3-7421-4ab9-be1f-318db2853596 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.788251761 +0000 UTC m=+2.791445802 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert") pod "ingress-canary-knq92" (UID: "b7508cd3-7421-4ab9-be1f-318db2853596") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.288326 master-0 kubenswrapper[24928]: E1205 10:47:22.288285 24928 secret.go:189] Couldn't get secret openshift-cluster-machine-approver/machine-approver-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.288398 master-0 kubenswrapper[24928]: E1205 10:47:22.288374 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/62f174ff-9f3c-4051-a34a-e55a59207171-machine-approver-tls podName:62f174ff-9f3c-4051-a34a-e55a59207171 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.788354904 +0000 UTC m=+2.791548785 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-approver-tls" (UniqueName: "kubernetes.io/secret/62f174ff-9f3c-4051-a34a-e55a59207171-machine-approver-tls") pod "machine-approver-74d9cbffbc-9jbnk" (UID: "62f174ff-9f3c-4051-a34a-e55a59207171") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.289685 master-0 kubenswrapper[24928]: E1205 10:47:22.289647 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/metrics-client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.289746 master-0 kubenswrapper[24928]: E1205 10:47:22.289712 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/21a9bd5a-3ade-48ef-8004-e0492daa85ba-metrics-client-ca podName:21a9bd5a-3ade-48ef-8004-e0492daa85ba nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.789694486 +0000 UTC m=+2.792888377 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-client-ca" (UniqueName: "kubernetes.io/configmap/21a9bd5a-3ade-48ef-8004-e0492daa85ba-metrics-client-ca") pod "node-exporter-bmqsb" (UID: "21a9bd5a-3ade-48ef-8004-e0492daa85ba") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.289746 master-0 kubenswrapper[24928]: E1205 10:47:22.289708 24928 configmap.go:193] Couldn't get configMap openshift-cloud-credential-operator/cco-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.289844 master-0 kubenswrapper[24928]: E1205 10:47:22.289725 24928 secret.go:189] Couldn't get secret openshift-machine-api/cluster-baremetal-webhook-server-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.289844 master-0 kubenswrapper[24928]: E1205 10:47:22.289817 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cco-trusted-ca podName:ee0b4a7f-1b96-4304-bef0-fb575e0e431c nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.789793459 +0000 UTC m=+2.792987450 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cco-trusted-ca" (UniqueName: "kubernetes.io/configmap/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cco-trusted-ca") pod "cloud-credential-operator-698c598cfc-rgc4p" (UID: "ee0b4a7f-1b96-4304-bef0-fb575e0e431c") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.289922 master-0 kubenswrapper[24928]: E1205 10:47:22.289853 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cert podName:48bd1d86-a6f2-439f-ab04-6a9a442bec42 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.78983799 +0000 UTC m=+2.793032011 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cert") pod "cluster-baremetal-operator-78f758c7b9-6t2gm" (UID: "48bd1d86-a6f2-439f-ab04-6a9a442bec42") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.290898 master-0 kubenswrapper[24928]: E1205 10:47:22.290853 24928 secret.go:189] Couldn't get secret openshift-monitoring/telemeter-client: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.290984 master-0 kubenswrapper[24928]: E1205 10:47:22.290916 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.790902306 +0000 UTC m=+2.794096157 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-telemeter-client" (UniqueName: "kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.290984 master-0 kubenswrapper[24928]: E1205 10:47:22.290945 24928 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.290984 master-0 kubenswrapper[24928]: E1205 10:47:22.290942 24928 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/packageserver-service-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.290984 master-0 kubenswrapper[24928]: E1205 10:47:22.290977 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs podName:3332b604-6f0d-4243-bce1-b3ab76e58c78 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.790968507 +0000 UTC m=+2.794162458 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs") pod "multus-admission-controller-8dbbb5754-7p9c2" (UID: "3332b604-6f0d-4243-bce1-b3ab76e58c78") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.291180 master-0 kubenswrapper[24928]: E1205 10:47:22.291018 24928 secret.go:189] Couldn't get secret openshift-monitoring/node-exporter-kube-rbac-proxy-config: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.291180 master-0 kubenswrapper[24928]: E1205 10:47:22.291034 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-apiservice-cert podName:2c3fab4d-05b8-42ec-8c91-91ff64562649 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.791011358 +0000 UTC m=+2.794205349 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-apiservice-cert") pod "packageserver-d7b67d8cf-krp6c" (UID: "2c3fab4d-05b8-42ec-8c91-91ff64562649") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.291180 master-0 kubenswrapper[24928]: E1205 10:47:22.291058 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-kube-rbac-proxy-config podName:21a9bd5a-3ade-48ef-8004-e0492daa85ba nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.791049289 +0000 UTC m=+2.794243340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-kube-rbac-proxy-config" (UniqueName: "kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-kube-rbac-proxy-config") pod "node-exporter-bmqsb" (UID: "21a9bd5a-3ade-48ef-8004-e0492daa85ba") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.291180 master-0 kubenswrapper[24928]: E1205 10:47:22.291138 24928 secret.go:189] Couldn't get secret openshift-monitoring/prometheus-operator-kube-rbac-proxy-config: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.291180 master-0 kubenswrapper[24928]: E1205 10:47:22.291173 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-kube-rbac-proxy-config podName:b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.791164272 +0000 UTC m=+2.794358203 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "prometheus-operator-kube-rbac-proxy-config" (UniqueName: "kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-kube-rbac-proxy-config") pod "prometheus-operator-6c74d9cb9f-r787z" (UID: "b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.292261 master-0 kubenswrapper[24928]: E1205 10:47:22.292216 24928 configmap.go:193] Couldn't get configMap openshift-cluster-machine-approver/machine-approver-config: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.292261 master-0 kubenswrapper[24928]: E1205 10:47:22.292240 24928 secret.go:189] Couldn't get secret openshift-machine-api/cluster-baremetal-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.292403 master-0 kubenswrapper[24928]: E1205 10:47:22.292292 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-config podName:62f174ff-9f3c-4051-a34a-e55a59207171 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.792271099 +0000 UTC m=+2.795464990 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-config") pod "machine-approver-74d9cbffbc-9jbnk" (UID: "62f174ff-9f3c-4051-a34a-e55a59207171") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.292403 master-0 kubenswrapper[24928]: E1205 10:47:22.292317 24928 secret.go:189] Couldn't get secret openshift-monitoring/openshift-state-metrics-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.292403 master-0 kubenswrapper[24928]: E1205 10:47:22.292328 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cluster-baremetal-operator-tls podName:48bd1d86-a6f2-439f-ab04-6a9a442bec42 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.79231207 +0000 UTC m=+2.795505951 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cluster-baremetal-operator-tls" (UniqueName: "kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cluster-baremetal-operator-tls") pod "cluster-baremetal-operator-78f758c7b9-6t2gm" (UID: "48bd1d86-a6f2-439f-ab04-6a9a442bec42") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.292403 master-0 kubenswrapper[24928]: E1205 10:47:22.292339 24928 secret.go:189] Couldn't get secret openshift-machine-api/cluster-autoscaler-operator-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.292403 master-0 kubenswrapper[24928]: E1205 10:47:22.292386 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-tls podName:d9ff1ae2-8547-42db-9fd0-f4782589ca18 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.792345821 +0000 UTC m=+2.795539712 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openshift-state-metrics-tls" (UniqueName: "kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-tls") pod "openshift-state-metrics-5974b6b869-9p5mt" (UID: "d9ff1ae2-8547-42db-9fd0-f4782589ca18") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.292403 master-0 kubenswrapper[24928]: E1205 10:47:22.292382 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/metrics-client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.293062 master-0 kubenswrapper[24928]: E1205 10:47:22.292417 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1f132702-304b-46ff-b428-6df1deeffec3-cert podName:1f132702-304b-46ff-b428-6df1deeffec3 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.792399952 +0000 UTC m=+2.795593843 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1f132702-304b-46ff-b428-6df1deeffec3-cert") pod "cluster-autoscaler-operator-5f49d774cd-cfg5f" (UID: "1f132702-304b-46ff-b428-6df1deeffec3") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.293062 master-0 kubenswrapper[24928]: E1205 10:47:22.292308 24928 configmap.go:193] Couldn't get configMap openshift-insights/service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.293062 master-0 kubenswrapper[24928]: E1205 10:47:22.292444 24928 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.293062 master-0 kubenswrapper[24928]: E1205 10:47:22.292511 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d9ff1ae2-8547-42db-9fd0-f4782589ca18-metrics-client-ca podName:d9ff1ae2-8547-42db-9fd0-f4782589ca18 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.792484474 +0000 UTC m=+2.795678385 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-client-ca" (UniqueName: "kubernetes.io/configmap/d9ff1ae2-8547-42db-9fd0-f4782589ca18-metrics-client-ca") pod "openshift-state-metrics-5974b6b869-9p5mt" (UID: "d9ff1ae2-8547-42db-9fd0-f4782589ca18") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.293062 master-0 kubenswrapper[24928]: E1205 10:47:22.292541 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6728a458-098b-44f9-8c8f-b0d76c5825fa-mcc-auth-proxy-config podName:6728a458-098b-44f9-8c8f-b0d76c5825fa nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.792530545 +0000 UTC m=+2.795724436 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "mcc-auth-proxy-config" (UniqueName: "kubernetes.io/configmap/6728a458-098b-44f9-8c8f-b0d76c5825fa-mcc-auth-proxy-config") pod "machine-config-controller-7c6d64c4cd-blwfs" (UID: "6728a458-098b-44f9-8c8f-b0d76c5825fa") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.293062 master-0 kubenswrapper[24928]: E1205 10:47:22.292571 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-service-ca-bundle podName:65e06ca6-1f15-43ec-b8dc-a300e1cf83bd nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.792561496 +0000 UTC m=+2.795755377 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-service-ca-bundle") pod "insights-operator-55965856b6-2sxv7" (UID: "65e06ca6-1f15-43ec-b8dc-a300e1cf83bd") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.293547 master-0 kubenswrapper[24928]: I1205 10:47:22.293489 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 10:47:22.293659 master-0 kubenswrapper[24928]: E1205 10:47:22.293493 24928 secret.go:189] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.293804 master-0 kubenswrapper[24928]: E1205 10:47:22.293724 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/82464e92-4de2-48f0-8772-a489abb16898-machine-api-operator-tls podName:82464e92-4de2-48f0-8772-a489abb16898 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.793693034 +0000 UTC m=+2.796886915 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/82464e92-4de2-48f0-8772-a489abb16898-machine-api-operator-tls") pod "machine-api-operator-88d48b57d-x7jfs" (UID: "82464e92-4de2-48f0-8772-a489abb16898") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.294652 master-0 kubenswrapper[24928]: E1205 10:47:22.294623 24928 secret.go:189] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.294692 master-0 kubenswrapper[24928]: E1205 10:47:22.294665 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-certs podName:568e9bed-cbd2-49bd-84af-9e17f95a003f nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.794656998 +0000 UTC m=+2.797850839 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-certs") pod "machine-config-server-5t4nn" (UID: "568e9bed-cbd2-49bd-84af-9e17f95a003f") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.294771 master-0 kubenswrapper[24928]: E1205 10:47:22.294717 24928 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.294876 master-0 kubenswrapper[24928]: E1205 10:47:22.294844 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-config podName:82464e92-4de2-48f0-8772-a489abb16898 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.794818271 +0000 UTC m=+2.798012132 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-config") pod "machine-api-operator-88d48b57d-x7jfs" (UID: "82464e92-4de2-48f0-8772-a489abb16898") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.295089 master-0 kubenswrapper[24928]: E1205 10:47:22.295054 24928 secret.go:189] Couldn't get secret openshift-machine-config-operator/mcc-proxy-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.295152 master-0 kubenswrapper[24928]: E1205 10:47:22.295131 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6728a458-098b-44f9-8c8f-b0d76c5825fa-proxy-tls podName:6728a458-098b-44f9-8c8f-b0d76c5825fa nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.795114688 +0000 UTC m=+2.798308579 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/6728a458-098b-44f9-8c8f-b0d76c5825fa-proxy-tls") pod "machine-config-controller-7c6d64c4cd-blwfs" (UID: "6728a458-098b-44f9-8c8f-b0d76c5825fa") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.296170 master-0 kubenswrapper[24928]: E1205 10:47:22.296135 24928 secret.go:189] Couldn't get secret openshift-machine-api/control-plane-machine-set-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.296217 master-0 kubenswrapper[24928]: E1205 10:47:22.296175 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-control-plane-machine-set-operator-tls podName:24444474-aa9a-4a0f-8b4d-90f0009e0dc7 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.796166994 +0000 UTC m=+2.799360845 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "control-plane-machine-set-operator-tls" (UniqueName: "kubernetes.io/secret/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-control-plane-machine-set-operator-tls") pod "control-plane-machine-set-operator-7df95c79b5-qnq6t" (UID: "24444474-aa9a-4a0f-8b4d-90f0009e0dc7") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.296217 master-0 kubenswrapper[24928]: E1205 10:47:22.296177 24928 secret.go:189] Couldn't get secret openshift-cloud-credential-operator/cloud-credential-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.296217 master-0 kubenswrapper[24928]: E1205 10:47:22.296206 24928 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.296320 master-0 kubenswrapper[24928]: E1205 10:47:22.296215 24928 secret.go:189] Couldn't get secret openshift-monitoring/federate-client-certs: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.296320 master-0 kubenswrapper[24928]: E1205 10:47:22.296257 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/telemeter-trusted-ca-bundle-56c9b9fa8d9gs: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.296320 master-0 kubenswrapper[24928]: E1205 10:47:22.296227 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-images podName:82464e92-4de2-48f0-8772-a489abb16898 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.796221726 +0000 UTC m=+2.799415577 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-images") pod "machine-api-operator-88d48b57d-x7jfs" (UID: "82464e92-4de2-48f0-8772-a489abb16898") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.296320 master-0 kubenswrapper[24928]: E1205 10:47:22.296299 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cloud-credential-operator-serving-cert podName:ee0b4a7f-1b96-4304-bef0-fb575e0e431c nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.796280037 +0000 UTC m=+2.799473918 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cloud-credential-operator-serving-cert" (UniqueName: "kubernetes.io/secret/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cloud-credential-operator-serving-cert") pod "cloud-credential-operator-698c598cfc-rgc4p" (UID: "ee0b4a7f-1b96-4304-bef0-fb575e0e431c") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.296454 master-0 kubenswrapper[24928]: E1205 10:47:22.296327 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.796314238 +0000 UTC m=+2.799508129 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "federate-client-tls" (UniqueName: "kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.296454 master-0 kubenswrapper[24928]: E1205 10:47:22.296350 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.796340068 +0000 UTC m=+2.799533959 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "telemeter-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.297478 master-0 kubenswrapper[24928]: E1205 10:47:22.297450 24928 secret.go:189] Couldn't get secret openshift-monitoring/kube-state-metrics-kube-rbac-proxy-config: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.297478 master-0 kubenswrapper[24928]: E1205 10:47:22.297467 24928 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.297554 master-0 kubenswrapper[24928]: E1205 10:47:22.297488 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-kube-rbac-proxy-config podName:54f9f9d4-8c31-4636-8010-22ee11b9b323 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.797480476 +0000 UTC m=+2.800674327 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-state-metrics-kube-rbac-proxy-config" (UniqueName: "kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-kube-rbac-proxy-config") pod "kube-state-metrics-5857974f64-xj7pj" (UID: "54f9f9d4-8c31-4636-8010-22ee11b9b323") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.297554 master-0 kubenswrapper[24928]: E1205 10:47:22.297510 24928 secret.go:189] Couldn't get secret openshift-monitoring/openshift-state-metrics-kube-rbac-proxy-config: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.297554 master-0 kubenswrapper[24928]: E1205 10:47:22.297527 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-images podName:01488ac4-313d-48d5-9e86-7b2011b9e91e nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.797499917 +0000 UTC m=+2.800693768 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-images") pod "machine-config-operator-dc5d7666f-2cf9h" (UID: "01488ac4-313d-48d5-9e86-7b2011b9e91e") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.297554 master-0 kubenswrapper[24928]: E1205 10:47:22.297539 24928 secret.go:189] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.297674 master-0 kubenswrapper[24928]: E1205 10:47:22.297545 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-kube-rbac-proxy-config podName:d9ff1ae2-8547-42db-9fd0-f4782589ca18 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.797537008 +0000 UTC m=+2.800730859 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "openshift-state-metrics-kube-rbac-proxy-config" (UniqueName: "kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-kube-rbac-proxy-config") pod "openshift-state-metrics-5974b6b869-9p5mt" (UID: "d9ff1ae2-8547-42db-9fd0-f4782589ca18") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.297674 master-0 kubenswrapper[24928]: E1205 10:47:22.297625 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/01488ac4-313d-48d5-9e86-7b2011b9e91e-proxy-tls podName:01488ac4-313d-48d5-9e86-7b2011b9e91e nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.797607199 +0000 UTC m=+2.800801090 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/01488ac4-313d-48d5-9e86-7b2011b9e91e-proxy-tls") pod "machine-config-operator-dc5d7666f-2cf9h" (UID: "01488ac4-313d-48d5-9e86-7b2011b9e91e") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.298673 master-0 kubenswrapper[24928]: E1205 10:47:22.298639 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/kube-state-metrics-custom-resource-state-configmap: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.298720 master-0 kubenswrapper[24928]: E1205 10:47:22.298685 24928 secret.go:189] Couldn't get secret openshift-monitoring/prometheus-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.298720 master-0 kubenswrapper[24928]: E1205 10:47:22.298689 24928 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy-cluster-autoscaler-operator: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.298780 master-0 kubenswrapper[24928]: E1205 10:47:22.298696 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-custom-resource-state-configmap podName:54f9f9d4-8c31-4636-8010-22ee11b9b323 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.798685125 +0000 UTC m=+2.801878976 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-state-metrics-custom-resource-state-configmap" (UniqueName: "kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-custom-resource-state-configmap") pod "kube-state-metrics-5857974f64-xj7pj" (UID: "54f9f9d4-8c31-4636-8010-22ee11b9b323") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.298780 master-0 kubenswrapper[24928]: E1205 10:47:22.298731 24928 secret.go:189] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.298780 master-0 kubenswrapper[24928]: E1205 10:47:22.298745 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls podName:b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.798733156 +0000 UTC m=+2.801927007 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "prometheus-operator-tls" (UniqueName: "kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls") pod "prometheus-operator-6c74d9cb9f-r787z" (UID: "b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.298780 master-0 kubenswrapper[24928]: E1205 10:47:22.298748 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/kubelet-serving-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.298780 master-0 kubenswrapper[24928]: E1205 10:47:22.298766 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1f132702-304b-46ff-b428-6df1deeffec3-auth-proxy-config podName:1f132702-304b-46ff-b428-6df1deeffec3 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.798757746 +0000 UTC m=+2.801951707 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "auth-proxy-config" (UniqueName: "kubernetes.io/configmap/1f132702-304b-46ff-b428-6df1deeffec3-auth-proxy-config") pod "cluster-autoscaler-operator-5f49d774cd-cfg5f" (UID: "1f132702-304b-46ff-b428-6df1deeffec3") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.298780 master-0 kubenswrapper[24928]: E1205 10:47:22.298784 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0de26264-46c2-4d29-97da-25a1681d6a8e-samples-operator-tls podName:0de26264-46c2-4d29-97da-25a1681d6a8e nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.798777257 +0000 UTC m=+2.801971108 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/0de26264-46c2-4d29-97da-25a1681d6a8e-samples-operator-tls") pod "cluster-samples-operator-797cfd8b47-glpx7" (UID: "0de26264-46c2-4d29-97da-25a1681d6a8e") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.298955 master-0 kubenswrapper[24928]: E1205 10:47:22.298828 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle podName:3aa9a063-322b-4dc6-a724-05a66026160b nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.798815488 +0000 UTC m=+2.802009339 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "configmap-kubelet-serving-ca-bundle" (UniqueName: "kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle") pod "metrics-server-7c46d76dff-z8d8z" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.301028 master-0 kubenswrapper[24928]: E1205 10:47:22.300988 24928 configmap.go:193] Couldn't get configMap openshift-cloud-controller-manager-operator/cloud-controller-manager-images: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.301099 master-0 kubenswrapper[24928]: E1205 10:47:22.301073 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-images podName:af7078c2-ad4f-415b-ba2f-77f5145c3b3f nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.801055193 +0000 UTC m=+2.804249084 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-images") pod "cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" (UID: "af7078c2-ad4f-415b-ba2f-77f5145c3b3f") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.301099 master-0 kubenswrapper[24928]: E1205 10:47:22.301077 24928 secret.go:189] Couldn't get secret openshift-cloud-controller-manager-operator/cloud-controller-manager-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.301169 master-0 kubenswrapper[24928]: E1205 10:47:22.301125 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-cloud-controller-manager-operator-tls podName:af7078c2-ad4f-415b-ba2f-77f5145c3b3f nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.801115344 +0000 UTC m=+2.804309195 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cloud-controller-manager-operator-tls" (UniqueName: "kubernetes.io/secret/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-cloud-controller-manager-operator-tls") pod "cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" (UID: "af7078c2-ad4f-415b-ba2f-77f5145c3b3f") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303191 master-0 kubenswrapper[24928]: E1205 10:47:22.303151 24928 secret.go:189] Couldn't get secret openshift-monitoring/telemeter-client-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303191 master-0 kubenswrapper[24928]: E1205 10:47:22.303189 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/metrics-client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.303300 master-0 kubenswrapper[24928]: E1205 10:47:22.303214 24928 secret.go:189] Couldn't get secret openshift-cluster-storage-operator/cluster-storage-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303300 master-0 kubenswrapper[24928]: E1205 10:47:22.303221 24928 secret.go:189] Couldn't get secret openshift-monitoring/metrics-server-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303300 master-0 kubenswrapper[24928]: E1205 10:47:22.303193 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.803184795 +0000 UTC m=+2.806378646 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "telemeter-client-tls" (UniqueName: "kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303300 master-0 kubenswrapper[24928]: E1205 10:47:22.303243 24928 secret.go:189] Couldn't get secret openshift-operator-lifecycle-manager/catalog-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303300 master-0 kubenswrapper[24928]: E1205 10:47:22.303248 24928 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.303300 master-0 kubenswrapper[24928]: E1205 10:47:22.303263 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-metrics-client-ca podName:54f9f9d4-8c31-4636-8010-22ee11b9b323 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.803249137 +0000 UTC m=+2.806443178 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-client-ca" (UniqueName: "kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-metrics-client-ca") pod "kube-state-metrics-5857974f64-xj7pj" (UID: "54f9f9d4-8c31-4636-8010-22ee11b9b323") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.303300 master-0 kubenswrapper[24928]: E1205 10:47:22.303279 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/58b95ae4-7f74-4777-8441-0a0ae28199e9-cluster-storage-operator-serving-cert podName:58b95ae4-7f74-4777-8441-0a0ae28199e9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.803271117 +0000 UTC m=+2.806465168 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cluster-storage-operator-serving-cert" (UniqueName: "kubernetes.io/secret/58b95ae4-7f74-4777-8441-0a0ae28199e9-cluster-storage-operator-serving-cert") pod "cluster-storage-operator-dcf7fc84b-9rzps" (UID: "58b95ae4-7f74-4777-8441-0a0ae28199e9") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303300 master-0 kubenswrapper[24928]: E1205 10:47:22.303284 24928 secret.go:189] Couldn't get secret openshift-monitoring/kube-state-metrics-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303300 master-0 kubenswrapper[24928]: E1205 10:47:22.303296 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls podName:3aa9a063-322b-4dc6-a724-05a66026160b nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.803288657 +0000 UTC m=+2.806482588 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-metrics-server-tls" (UniqueName: "kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls") pod "metrics-server-7c46d76dff-z8d8z" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303300 master-0 kubenswrapper[24928]: E1205 10:47:22.303302 24928 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.303300 master-0 kubenswrapper[24928]: E1205 10:47:22.303308 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/metrics-server-audit-profiles: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.303748 master-0 kubenswrapper[24928]: E1205 10:47:22.303313 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-srv-cert podName:4c2975ec-e33d-4960-a708-277d41c79b15 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.803305918 +0000 UTC m=+2.806499839 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-srv-cert") pod "catalog-operator-fbc6455c4-mbm77" (UID: "4c2975ec-e33d-4960-a708-277d41c79b15") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303748 master-0 kubenswrapper[24928]: E1205 10:47:22.303340 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-mcd-auth-proxy-config podName:b7e7f216-f9af-41df-a0b0-df2a76b9b72a nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.803330768 +0000 UTC m=+2.806524819 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "mcd-auth-proxy-config" (UniqueName: "kubernetes.io/configmap/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-mcd-auth-proxy-config") pod "machine-config-daemon-5n6nw" (UID: "b7e7f216-f9af-41df-a0b0-df2a76b9b72a") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.303748 master-0 kubenswrapper[24928]: E1205 10:47:22.303304 24928 secret.go:189] Couldn't get secret openshift-insights/openshift-insights-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303748 master-0 kubenswrapper[24928]: E1205 10:47:22.303356 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-tls podName:54f9f9d4-8c31-4636-8010-22ee11b9b323 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.803349559 +0000 UTC m=+2.806543620 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-state-metrics-tls" (UniqueName: "kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-tls") pod "kube-state-metrics-5857974f64-xj7pj" (UID: "54f9f9d4-8c31-4636-8010-22ee11b9b323") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303748 master-0 kubenswrapper[24928]: E1205 10:47:22.303372 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-auth-proxy-config podName:01488ac4-313d-48d5-9e86-7b2011b9e91e nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.803367889 +0000 UTC m=+2.806561740 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "auth-proxy-config" (UniqueName: "kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-auth-proxy-config") pod "machine-config-operator-dc5d7666f-2cf9h" (UID: "01488ac4-313d-48d5-9e86-7b2011b9e91e") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.303748 master-0 kubenswrapper[24928]: E1205 10:47:22.303356 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/telemeter-client-serving-certs-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.303748 master-0 kubenswrapper[24928]: E1205 10:47:22.303416 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles podName:3aa9a063-322b-4dc6-a724-05a66026160b nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.80337828 +0000 UTC m=+2.806572131 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-server-audit-profiles" (UniqueName: "kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles") pod "metrics-server-7c46d76dff-z8d8z" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.303748 master-0 kubenswrapper[24928]: E1205 10:47:22.303466 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-serving-cert podName:65e06ca6-1f15-43ec-b8dc-a300e1cf83bd nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.803459891 +0000 UTC m=+2.806653733 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-serving-cert") pod "insights-operator-55965856b6-2sxv7" (UID: "65e06ca6-1f15-43ec-b8dc-a300e1cf83bd") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.303748 master-0 kubenswrapper[24928]: E1205 10:47:22.303484 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.803479022 +0000 UTC m=+2.806672873 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-certs-ca-bundle" (UniqueName: "kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:22.304605 master-0 kubenswrapper[24928]: E1205 10:47:22.304571 24928 secret.go:189] Couldn't get secret openshift-machine-config-operator/proxy-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.304668 master-0 kubenswrapper[24928]: E1205 10:47:22.304651 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-proxy-tls podName:b7e7f216-f9af-41df-a0b0-df2a76b9b72a nodeName:}" failed. No retries permitted until 2025-12-05 10:47:22.80463678 +0000 UTC m=+2.807830621 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-proxy-tls") pod "machine-config-daemon-5n6nw" (UID: "b7e7f216-f9af-41df-a0b0-df2a76b9b72a") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:22.312867 master-0 kubenswrapper[24928]: I1205 10:47:22.312837 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-pz2dt" Dec 05 10:47:22.333767 master-0 kubenswrapper[24928]: I1205 10:47:22.333666 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 10:47:22.353334 master-0 kubenswrapper[24928]: I1205 10:47:22.353257 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 10:47:22.373343 master-0 kubenswrapper[24928]: I1205 10:47:22.373247 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 10:47:22.392740 master-0 kubenswrapper[24928]: I1205 10:47:22.392678 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 10:47:22.412573 master-0 kubenswrapper[24928]: I1205 10:47:22.412496 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-kfn9l" Dec 05 10:47:22.434071 master-0 kubenswrapper[24928]: I1205 10:47:22.433977 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"kube-root-ca.crt" Dec 05 10:47:22.453234 master-0 kubenswrapper[24928]: I1205 10:47:22.453138 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"service-ca-bundle" Dec 05 10:47:22.473148 master-0 kubenswrapper[24928]: I1205 10:47:22.472952 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:22.473148 master-0 kubenswrapper[24928]: I1205 10:47:22.473110 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:22.475245 master-0 kubenswrapper[24928]: I1205 10:47:22.474794 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 10:47:22.493314 master-0 kubenswrapper[24928]: I1205 10:47:22.493273 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-insights"/"openshift-insights-serving-cert" Dec 05 10:47:22.512664 master-0 kubenswrapper[24928]: I1205 10:47:22.512602 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-insights"/"operator-dockercfg-lngmk" Dec 05 10:47:22.532307 master-0 kubenswrapper[24928]: I1205 10:47:22.532183 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-credential-operator"/"openshift-service-ca.crt" Dec 05 10:47:22.553545 master-0 kubenswrapper[24928]: I1205 10:47:22.553460 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"openshift-service-ca.crt" Dec 05 10:47:22.572312 master-0 kubenswrapper[24928]: I1205 10:47:22.572255 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-credential-operator"/"cloud-credential-operator-serving-cert" Dec 05 10:47:22.592975 master-0 kubenswrapper[24928]: I1205 10:47:22.592935 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-credential-operator"/"cloud-credential-operator-dockercfg-ftghr" Dec 05 10:47:22.619651 master-0 kubenswrapper[24928]: I1205 10:47:22.619565 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-credential-operator"/"cco-trusted-ca" Dec 05 10:47:22.632648 master-0 kubenswrapper[24928]: I1205 10:47:22.632605 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 10:47:22.662535 master-0 kubenswrapper[24928]: I1205 10:47:22.662472 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"trusted-ca-bundle" Dec 05 10:47:22.706785 master-0 kubenswrapper[24928]: I1205 10:47:22.706723 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-credential-operator"/"kube-root-ca.crt" Dec 05 10:47:22.708248 master-0 kubenswrapper[24928]: I1205 10:47:22.708215 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-9zg4x" Dec 05 10:47:22.713115 master-0 kubenswrapper[24928]: I1205 10:47:22.713069 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 10:47:22.734342 master-0 kubenswrapper[24928]: I1205 10:47:22.734251 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 10:47:22.752440 master-0 kubenswrapper[24928]: I1205 10:47:22.752379 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-storage-operator"/"cluster-storage-operator-dockercfg-g78vb" Dec 05 10:47:22.773135 master-0 kubenswrapper[24928]: I1205 10:47:22.773043 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-storage-operator"/"cluster-storage-operator-serving-cert" Dec 05 10:47:22.792728 master-0 kubenswrapper[24928]: I1205 10:47:22.792664 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 10:47:22.812983 master-0 kubenswrapper[24928]: I1205 10:47:22.812905 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 10:47:22.833753 master-0 kubenswrapper[24928]: I1205 10:47:22.833679 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 10:47:22.853174 master-0 kubenswrapper[24928]: I1205 10:47:22.853099 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rrnj5" Dec 05 10:47:22.865356 master-0 kubenswrapper[24928]: I1205 10:47:22.865255 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:22.865660 master-0 kubenswrapper[24928]: I1205 10:47:22.865581 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:47:22.865849 master-0 kubenswrapper[24928]: I1205 10:47:22.865805 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:47:22.866177 master-0 kubenswrapper[24928]: I1205 10:47:22.866104 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-baremetal-operator-tls\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cluster-baremetal-operator-tls\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:22.866265 master-0 kubenswrapper[24928]: I1205 10:47:22.866220 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-tls\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:47:22.866347 master-0 kubenswrapper[24928]: I1205 10:47:22.866274 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d9ff1ae2-8547-42db-9fd0-f4782589ca18-metrics-client-ca\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:47:22.866483 master-0 kubenswrapper[24928]: I1205 10:47:22.866444 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6728a458-098b-44f9-8c8f-b0d76c5825fa-mcc-auth-proxy-config\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:47:22.866547 master-0 kubenswrapper[24928]: I1205 10:47:22.866514 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:47:22.866597 master-0 kubenswrapper[24928]: I1205 10:47:22.866553 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f132702-304b-46ff-b428-6df1deeffec3-cert\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:47:22.866641 master-0 kubenswrapper[24928]: I1205 10:47:22.866594 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-service-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:22.866756 master-0 kubenswrapper[24928]: I1205 10:47:22.866713 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/82464e92-4de2-48f0-8772-a489abb16898-machine-api-operator-tls\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:47:22.866968 master-0 kubenswrapper[24928]: I1205 10:47:22.866906 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-certs\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:47:22.867187 master-0 kubenswrapper[24928]: I1205 10:47:22.867116 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-config\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:47:22.867187 master-0 kubenswrapper[24928]: I1205 10:47:22.867159 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-service-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:22.867589 master-0 kubenswrapper[24928]: I1205 10:47:22.867269 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6728a458-098b-44f9-8c8f-b0d76c5825fa-proxy-tls\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:47:22.867589 master-0 kubenswrapper[24928]: I1205 10:47:22.867318 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-images\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:47:22.867589 master-0 kubenswrapper[24928]: I1205 10:47:22.867381 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-7df95c79b5-qnq6t\" (UID: \"24444474-aa9a-4a0f-8b4d-90f0009e0dc7\") " pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:47:22.867589 master-0 kubenswrapper[24928]: I1205 10:47:22.867470 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:22.867589 master-0 kubenswrapper[24928]: I1205 10:47:22.867487 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6728a458-098b-44f9-8c8f-b0d76c5825fa-mcc-auth-proxy-config\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:47:22.868156 master-0 kubenswrapper[24928]: I1205 10:47:22.867631 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:22.868156 master-0 kubenswrapper[24928]: I1205 10:47:22.867712 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloud-credential-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cloud-credential-operator-serving-cert\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:47:22.868156 master-0 kubenswrapper[24928]: I1205 10:47:22.867852 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:22.868156 master-0 kubenswrapper[24928]: I1205 10:47:22.867927 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-7df95c79b5-qnq6t\" (UID: \"24444474-aa9a-4a0f-8b4d-90f0009e0dc7\") " pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:47:22.868156 master-0 kubenswrapper[24928]: I1205 10:47:22.867945 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:47:22.868156 master-0 kubenswrapper[24928]: I1205 10:47:22.868065 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/01488ac4-313d-48d5-9e86-7b2011b9e91e-proxy-tls\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:47:22.868156 master-0 kubenswrapper[24928]: I1205 10:47:22.868129 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-images\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:47:22.868848 master-0 kubenswrapper[24928]: I1205 10:47:22.868215 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloud-credential-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cloud-credential-operator-serving-cert\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:47:22.868848 master-0 kubenswrapper[24928]: I1205 10:47:22.868342 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:22.868848 master-0 kubenswrapper[24928]: I1205 10:47:22.868464 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0de26264-46c2-4d29-97da-25a1681d6a8e-samples-operator-tls\") pod \"cluster-samples-operator-797cfd8b47-glpx7\" (UID: \"0de26264-46c2-4d29-97da-25a1681d6a8e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:47:22.868848 master-0 kubenswrapper[24928]: I1205 10:47:22.868519 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/01488ac4-313d-48d5-9e86-7b2011b9e91e-proxy-tls\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:47:22.868848 master-0 kubenswrapper[24928]: I1205 10:47:22.868567 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:22.868848 master-0 kubenswrapper[24928]: I1205 10:47:22.868673 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f132702-304b-46ff-b428-6df1deeffec3-auth-proxy-config\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:47:22.868848 master-0 kubenswrapper[24928]: I1205 10:47:22.868682 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-images\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:47:22.868848 master-0 kubenswrapper[24928]: I1205 10:47:22.868768 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:47:22.868848 master-0 kubenswrapper[24928]: I1205 10:47:22.868811 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0de26264-46c2-4d29-97da-25a1681d6a8e-samples-operator-tls\") pod \"cluster-samples-operator-797cfd8b47-glpx7\" (UID: \"0de26264-46c2-4d29-97da-25a1681d6a8e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:47:22.869541 master-0 kubenswrapper[24928]: I1205 10:47:22.868869 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cloud-controller-manager-operator-tls\" (UniqueName: \"kubernetes.io/secret/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-cloud-controller-manager-operator-tls\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:22.869541 master-0 kubenswrapper[24928]: I1205 10:47:22.868939 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-proxy-tls\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:47:22.869541 master-0 kubenswrapper[24928]: I1205 10:47:22.868989 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-metrics-client-ca\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:22.869541 master-0 kubenswrapper[24928]: I1205 10:47:22.869109 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:22.869541 master-0 kubenswrapper[24928]: I1205 10:47:22.869197 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-serving-cert\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:22.869541 master-0 kubenswrapper[24928]: I1205 10:47:22.869291 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-srv-cert\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:47:22.869541 master-0 kubenswrapper[24928]: I1205 10:47:22.869350 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-images\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:22.869541 master-0 kubenswrapper[24928]: I1205 10:47:22.869457 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-storage-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/58b95ae4-7f74-4777-8441-0a0ae28199e9-cluster-storage-operator-serving-cert\") pod \"cluster-storage-operator-dcf7fc84b-9rzps\" (UID: \"58b95ae4-7f74-4777-8441-0a0ae28199e9\") " pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:47:22.869541 master-0 kubenswrapper[24928]: I1205 10:47:22.869522 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:22.870116 master-0 kubenswrapper[24928]: I1205 10:47:22.869566 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-mcd-auth-proxy-config\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:47:22.870116 master-0 kubenswrapper[24928]: I1205 10:47:22.869666 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-serving-cert\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:22.870116 master-0 kubenswrapper[24928]: I1205 10:47:22.869801 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-auth-proxy-config\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:47:22.870116 master-0 kubenswrapper[24928]: I1205 10:47:22.869847 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-mcd-auth-proxy-config\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:47:22.870116 master-0 kubenswrapper[24928]: I1205 10:47:22.869900 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-tls\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:22.870116 master-0 kubenswrapper[24928]: I1205 10:47:22.869910 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-storage-operator-serving-cert\" (UniqueName: \"kubernetes.io/secret/58b95ae4-7f74-4777-8441-0a0ae28199e9-cluster-storage-operator-serving-cert\") pod \"cluster-storage-operator-dcf7fc84b-9rzps\" (UID: \"58b95ae4-7f74-4777-8441-0a0ae28199e9\") " pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:47:22.870116 master-0 kubenswrapper[24928]: I1205 10:47:22.870002 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:22.870116 master-0 kubenswrapper[24928]: I1205 10:47:22.870063 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:22.870116 master-0 kubenswrapper[24928]: I1205 10:47:22.870075 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/01488ac4-313d-48d5-9e86-7b2011b9e91e-auth-proxy-config\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:47:22.870116 master-0 kubenswrapper[24928]: I1205 10:47:22.870107 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-node-bootstrap-token\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:47:22.870782 master-0 kubenswrapper[24928]: I1205 10:47:22.870148 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-metrics-client-ca\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:47:22.870782 master-0 kubenswrapper[24928]: I1205 10:47:22.870187 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-images\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:22.870782 master-0 kubenswrapper[24928]: I1205 10:47:22.870302 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-auth-proxy-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:47:22.870782 master-0 kubenswrapper[24928]: I1205 10:47:22.870350 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-auth-proxy-config\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:22.870782 master-0 kubenswrapper[24928]: I1205 10:47:22.870454 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-metrics-client-ca\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:22.870782 master-0 kubenswrapper[24928]: I1205 10:47:22.870493 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-trusted-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:22.870782 master-0 kubenswrapper[24928]: I1205 10:47:22.870568 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-config\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:22.870782 master-0 kubenswrapper[24928]: I1205 10:47:22.870746 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-tls\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:22.871492 master-0 kubenswrapper[24928]: I1205 10:47:22.870817 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:22.871492 master-0 kubenswrapper[24928]: I1205 10:47:22.870853 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-webhook-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:22.871492 master-0 kubenswrapper[24928]: I1205 10:47:22.870903 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:22.871492 master-0 kubenswrapper[24928]: I1205 10:47:22.871028 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-trusted-ca-bundle\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:22.871492 master-0 kubenswrapper[24928]: I1205 10:47:22.871093 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:22.871492 master-0 kubenswrapper[24928]: I1205 10:47:22.871130 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:47:22.871492 master-0 kubenswrapper[24928]: I1205 10:47:22.871203 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/62f174ff-9f3c-4051-a34a-e55a59207171-machine-approver-tls\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:47:22.871492 master-0 kubenswrapper[24928]: I1205 10:47:22.871407 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cco-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cco-trusted-ca\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:47:22.872227 master-0 kubenswrapper[24928]: I1205 10:47:22.871557 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cert\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:22.872227 master-0 kubenswrapper[24928]: I1205 10:47:22.871648 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/21a9bd5a-3ade-48ef-8004-e0492daa85ba-metrics-client-ca\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:22.872227 master-0 kubenswrapper[24928]: I1205 10:47:22.871787 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-apiservice-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:22.872227 master-0 kubenswrapper[24928]: I1205 10:47:22.871857 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:22.872227 master-0 kubenswrapper[24928]: I1205 10:47:22.872092 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cco-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-cco-trusted-ca\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:47:22.873346 master-0 kubenswrapper[24928]: I1205 10:47:22.873283 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 10:47:22.880554 master-0 kubenswrapper[24928]: I1205 10:47:22.880399 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4c2975ec-e33d-4960-a708-277d41c79b15-srv-cert\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:47:22.892473 master-0 kubenswrapper[24928]: I1205 10:47:22.892397 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-m8wb4" Dec 05 10:47:22.913789 master-0 kubenswrapper[24928]: I1205 10:47:22.913708 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 10:47:22.917628 master-0 kubenswrapper[24928]: I1205 10:47:22.917569 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-config\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:47:22.932708 master-0 kubenswrapper[24928]: I1205 10:47:22.932639 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 10:47:22.937169 master-0 kubenswrapper[24928]: I1205 10:47:22.937125 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/82464e92-4de2-48f0-8772-a489abb16898-machine-api-operator-tls\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:47:22.961289 master-0 kubenswrapper[24928]: I1205 10:47:22.956888 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-autoscaler-operator-dockercfg-vwmqb" Dec 05 10:47:22.978548 master-0 kubenswrapper[24928]: I1205 10:47:22.975462 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 10:47:22.978762 master-0 kubenswrapper[24928]: I1205 10:47:22.978606 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/82464e92-4de2-48f0-8772-a489abb16898-images\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:47:22.992435 master-0 kubenswrapper[24928]: I1205 10:47:22.992332 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-baremetal-webhook-server-cert" Dec 05 10:47:23.002480 master-0 kubenswrapper[24928]: I1205 10:47:23.002377 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cert\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:23.012846 master-0 kubenswrapper[24928]: I1205 10:47:23.012790 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-autoscaler-operator-cert" Dec 05 10:47:23.017407 master-0 kubenswrapper[24928]: I1205 10:47:23.017350 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1f132702-304b-46ff-b428-6df1deeffec3-cert\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:47:23.032724 master-0 kubenswrapper[24928]: I1205 10:47:23.032668 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy-cluster-autoscaler-operator" Dec 05 10:47:23.040040 master-0 kubenswrapper[24928]: I1205 10:47:23.039999 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1f132702-304b-46ff-b428-6df1deeffec3-auth-proxy-config\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:47:23.052641 master-0 kubenswrapper[24928]: I1205 10:47:23.052506 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-baremetal-operator-dockercfg-h7kbf" Dec 05 10:47:23.072113 master-0 kubenswrapper[24928]: I1205 10:47:23.072067 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-baremetal-operator-tls" Dec 05 10:47:23.077014 master-0 kubenswrapper[24928]: I1205 10:47:23.076976 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-baremetal-operator-tls\" (UniqueName: \"kubernetes.io/secret/48bd1d86-a6f2-439f-ab04-6a9a442bec42-cluster-baremetal-operator-tls\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:23.092160 master-0 kubenswrapper[24928]: I1205 10:47:23.092095 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"cluster-baremetal-operator-images" Dec 05 10:47:23.101007 master-0 kubenswrapper[24928]: I1205 10:47:23.100923 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-images\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:23.113005 master-0 kubenswrapper[24928]: I1205 10:47:23.112948 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"baremetal-kube-rbac-proxy" Dec 05 10:47:23.121025 master-0 kubenswrapper[24928]: I1205 10:47:23.120968 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48bd1d86-a6f2-439f-ab04-6a9a442bec42-config\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:23.132437 master-0 kubenswrapper[24928]: I1205 10:47:23.132355 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 10:47:23.142324 master-0 kubenswrapper[24928]: I1205 10:47:23.142263 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-apiservice-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:23.142510 master-0 kubenswrapper[24928]: I1205 10:47:23.142407 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2c3fab4d-05b8-42ec-8c91-91ff64562649-webhook-cert\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:23.152709 master-0 kubenswrapper[24928]: I1205 10:47:23.152642 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 10:47:23.172217 master-0 kubenswrapper[24928]: I1205 10:47:23.172151 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-2qdv2" Dec 05 10:47:23.191150 master-0 kubenswrapper[24928]: I1205 10:47:23.191108 24928 request.go:700] Waited for 1.995306493s due to client-side throttling, not priority and fairness, request: GET:https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-cluster-machine-approver/secrets?fieldSelector=metadata.name%3Dmachine-approver-tls&limit=500&resourceVersion=0 Dec 05 10:47:23.192979 master-0 kubenswrapper[24928]: I1205 10:47:23.192929 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 10:47:23.201974 master-0 kubenswrapper[24928]: I1205 10:47:23.201912 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/62f174ff-9f3c-4051-a34a-e55a59207171-machine-approver-tls\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:47:23.213104 master-0 kubenswrapper[24928]: I1205 10:47:23.213035 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 10:47:23.233520 master-0 kubenswrapper[24928]: I1205 10:47:23.233443 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 10:47:23.240931 master-0 kubenswrapper[24928]: I1205 10:47:23.240889 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:47:23.241051 master-0 kubenswrapper[24928]: I1205 10:47:23.241005 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-auth-proxy-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:47:23.241090 master-0 kubenswrapper[24928]: I1205 10:47:23.241068 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:23.245541 master-0 kubenswrapper[24928]: I1205 10:47:23.245501 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:47:23.252906 master-0 kubenswrapper[24928]: I1205 10:47:23.252860 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 10:47:23.257336 master-0 kubenswrapper[24928]: I1205 10:47:23.257299 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62f174ff-9f3c-4051-a34a-e55a59207171-config\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:47:23.272561 master-0 kubenswrapper[24928]: I1205 10:47:23.272492 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-fvzgd" Dec 05 10:47:23.291903 master-0 kubenswrapper[24928]: I1205 10:47:23.291834 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-vp7fw" Dec 05 10:47:23.297845 master-0 kubenswrapper[24928]: I1205 10:47:23.297762 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:23.312638 master-0 kubenswrapper[24928]: I1205 10:47:23.312492 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-6j2qd" Dec 05 10:47:23.332501 master-0 kubenswrapper[24928]: I1205 10:47:23.331605 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-qq76p" Dec 05 10:47:23.351936 master-0 kubenswrapper[24928]: I1205 10:47:23.351876 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 10:47:23.360031 master-0 kubenswrapper[24928]: I1205 10:47:23.359968 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-proxy-tls\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:47:23.373121 master-0 kubenswrapper[24928]: I1205 10:47:23.372921 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-s9q7m" Dec 05 10:47:23.392094 master-0 kubenswrapper[24928]: I1205 10:47:23.392004 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 10:47:23.413056 master-0 kubenswrapper[24928]: I1205 10:47:23.412858 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-controller-manager-operator"/"cluster-cloud-controller-manager-dockercfg-45kcs" Dec 05 10:47:23.442450 master-0 kubenswrapper[24928]: I1205 10:47:23.433242 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-controller-manager-operator"/"cloud-controller-manager-operator-tls" Dec 05 10:47:23.443174 master-0 kubenswrapper[24928]: I1205 10:47:23.443087 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cloud-controller-manager-operator-tls\" (UniqueName: \"kubernetes.io/secret/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-cloud-controller-manager-operator-tls\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:23.452663 master-0 kubenswrapper[24928]: I1205 10:47:23.452608 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"cloud-controller-manager-images" Dec 05 10:47:23.460707 master-0 kubenswrapper[24928]: I1205 10:47:23.460635 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-images\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:23.472206 master-0 kubenswrapper[24928]: I1205 10:47:23.472162 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"kube-rbac-proxy" Dec 05 10:47:23.481445 master-0 kubenswrapper[24928]: I1205 10:47:23.481381 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-auth-proxy-config\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:23.493278 master-0 kubenswrapper[24928]: I1205 10:47:23.493231 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:47:23.513007 master-0 kubenswrapper[24928]: I1205 10:47:23.512948 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 10:47:23.518190 master-0 kubenswrapper[24928]: I1205 10:47:23.518148 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/6728a458-098b-44f9-8c8f-b0d76c5825fa-proxy-tls\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:47:23.532447 master-0 kubenswrapper[24928]: I1205 10:47:23.532393 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-brtjg" Dec 05 10:47:23.553934 master-0 kubenswrapper[24928]: I1205 10:47:23.553875 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 10:47:23.560833 master-0 kubenswrapper[24928]: I1205 10:47:23.560791 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-node-bootstrap-token\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:47:23.573323 master-0 kubenswrapper[24928]: I1205 10:47:23.573163 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-zlzd6" Dec 05 10:47:23.592360 master-0 kubenswrapper[24928]: I1205 10:47:23.592291 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 10:47:23.597457 master-0 kubenswrapper[24928]: I1205 10:47:23.597364 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/568e9bed-cbd2-49bd-84af-9e17f95a003f-certs\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:47:23.612347 master-0 kubenswrapper[24928]: I1205 10:47:23.612279 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-2jmqj" Dec 05 10:47:23.634177 master-0 kubenswrapper[24928]: I1205 10:47:23.634117 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Dec 05 10:47:23.639606 master-0 kubenswrapper[24928]: I1205 10:47:23.639533 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-tls\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:47:23.652992 master-0 kubenswrapper[24928]: I1205 10:47:23.652901 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Dec 05 10:47:23.656657 master-0 kubenswrapper[24928]: I1205 10:47:23.656613 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:47:23.672761 master-0 kubenswrapper[24928]: I1205 10:47:23.672697 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-qkdkt" Dec 05 10:47:23.692714 master-0 kubenswrapper[24928]: I1205 10:47:23.692626 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Dec 05 10:47:23.700725 master-0 kubenswrapper[24928]: I1205 10:47:23.700659 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-tls\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:23.712722 master-0 kubenswrapper[24928]: I1205 10:47:23.712668 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Dec 05 10:47:23.718712 master-0 kubenswrapper[24928]: I1205 10:47:23.718657 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:23.737281 master-0 kubenswrapper[24928]: I1205 10:47:23.737228 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Dec 05 10:47:23.739725 master-0 kubenswrapper[24928]: I1205 10:47:23.739686 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:23.743126 master-0 kubenswrapper[24928]: I1205 10:47:23.743073 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-master-0" Dec 05 10:47:23.752109 master-0 kubenswrapper[24928]: I1205 10:47:23.752049 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Dec 05 10:47:23.757055 master-0 kubenswrapper[24928]: I1205 10:47:23.756973 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/d9ff1ae2-8547-42db-9fd0-f4782589ca18-metrics-client-ca\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:47:23.759378 master-0 kubenswrapper[24928]: I1205 10:47:23.759317 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/54f9f9d4-8c31-4636-8010-22ee11b9b323-metrics-client-ca\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:23.760574 master-0 kubenswrapper[24928]: I1205 10:47:23.760534 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-metrics-client-ca\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:47:23.760790 master-0 kubenswrapper[24928]: I1205 10:47:23.760753 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-metrics-client-ca\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:23.762858 master-0 kubenswrapper[24928]: I1205 10:47:23.762808 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/21a9bd5a-3ade-48ef-8004-e0492daa85ba-metrics-client-ca\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:23.774228 master-0 kubenswrapper[24928]: I1205 10:47:23.774056 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-x98dp" Dec 05 10:47:23.793269 master-0 kubenswrapper[24928]: I1205 10:47:23.793194 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Dec 05 10:47:23.796939 master-0 kubenswrapper[24928]: I1205 10:47:23.796881 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-tls\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:47:23.809780 master-0 kubenswrapper[24928]: I1205 10:47:23.809699 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:23.813239 master-0 kubenswrapper[24928]: I1205 10:47:23.813183 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Dec 05 10:47:23.816980 master-0 kubenswrapper[24928]: I1205 10:47:23.816916 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:23.819317 master-0 kubenswrapper[24928]: I1205 10:47:23.819291 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/d9ff1ae2-8547-42db-9fd0-f4782589ca18-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:47:23.831884 master-0 kubenswrapper[24928]: I1205 10:47:23.831748 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Dec 05 10:47:23.836089 master-0 kubenswrapper[24928]: I1205 10:47:23.836043 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:23.853864 master-0 kubenswrapper[24928]: I1205 10:47:23.853803 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Dec 05 10:47:23.862709 master-0 kubenswrapper[24928]: I1205 10:47:23.862654 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/21a9bd5a-3ade-48ef-8004-e0492daa85ba-node-exporter-tls\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:23.866440 master-0 kubenswrapper[24928]: E1205 10:47:23.866386 24928 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.866561 master-0 kubenswrapper[24928]: E1205 10:47:23.866472 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs podName:3332b604-6f0d-4243-bce1-b3ab76e58c78 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.866453543 +0000 UTC m=+4.869647394 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs") pod "multus-admission-controller-8dbbb5754-7p9c2" (UID: "3332b604-6f0d-4243-bce1-b3ab76e58c78") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.868621 master-0 kubenswrapper[24928]: E1205 10:47:23.868588 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/telemeter-trusted-ca-bundle-56c9b9fa8d9gs: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:23.868672 master-0 kubenswrapper[24928]: E1205 10:47:23.868631 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.868623086 +0000 UTC m=+4.871816937 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "telemeter-trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:23.868672 master-0 kubenswrapper[24928]: E1205 10:47:23.868635 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/kubelet-serving-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:23.868672 master-0 kubenswrapper[24928]: E1205 10:47:23.868648 24928 secret.go:189] Couldn't get secret openshift-monitoring/federate-client-certs: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.868672 master-0 kubenswrapper[24928]: E1205 10:47:23.868672 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.868667667 +0000 UTC m=+4.871861518 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "federate-client-tls" (UniqueName: "kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.868798 master-0 kubenswrapper[24928]: E1205 10:47:23.868696 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle podName:3aa9a063-322b-4dc6-a724-05a66026160b nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.868681988 +0000 UTC m=+4.871875849 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "configmap-kubelet-serving-ca-bundle" (UniqueName: "kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle") pod "metrics-server-7c46d76dff-z8d8z" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:23.869804 master-0 kubenswrapper[24928]: E1205 10:47:23.869776 24928 secret.go:189] Couldn't get secret openshift-monitoring/metrics-server-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.869857 master-0 kubenswrapper[24928]: E1205 10:47:23.869817 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls podName:3aa9a063-322b-4dc6-a724-05a66026160b nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.869808235 +0000 UTC m=+4.873002076 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-metrics-server-tls" (UniqueName: "kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls") pod "metrics-server-7c46d76dff-z8d8z" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.870118 master-0 kubenswrapper[24928]: E1205 10:47:23.870085 24928 secret.go:189] Couldn't get secret openshift-monitoring/telemeter-client-tls: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.870118 master-0 kubenswrapper[24928]: E1205 10:47:23.870102 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/metrics-server-audit-profiles: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:23.870208 master-0 kubenswrapper[24928]: E1205 10:47:23.870175 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.870131903 +0000 UTC m=+4.873325764 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "telemeter-client-tls" (UniqueName: "kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.870208 master-0 kubenswrapper[24928]: E1205 10:47:23.870203 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles podName:3aa9a063-322b-4dc6-a724-05a66026160b nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.870194804 +0000 UTC m=+4.873388665 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-server-audit-profiles" (UniqueName: "kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles") pod "metrics-server-7c46d76dff-z8d8z" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:23.871230 master-0 kubenswrapper[24928]: E1205 10:47:23.871191 24928 secret.go:189] Couldn't get secret openshift-monitoring/telemeter-client-kube-rbac-proxy-config: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.871307 master-0 kubenswrapper[24928]: E1205 10:47:23.871279 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.87126264 +0000 UTC m=+4.874456501 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-telemeter-client-kube-rbac-proxy-config" (UniqueName: "kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.871307 master-0 kubenswrapper[24928]: E1205 10:47:23.871196 24928 secret.go:189] Couldn't get secret openshift-monitoring/metrics-client-certs: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.871368 master-0 kubenswrapper[24928]: E1205 10:47:23.871300 24928 configmap.go:193] Couldn't get configMap openshift-monitoring/telemeter-client-serving-certs-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:23.871368 master-0 kubenswrapper[24928]: E1205 10:47:23.871344 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs podName:3aa9a063-322b-4dc6-a724-05a66026160b nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.871331922 +0000 UTC m=+4.874525803 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-metrics-client-certs" (UniqueName: "kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs") pod "metrics-server-7c46d76dff-z8d8z" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.871466 master-0 kubenswrapper[24928]: E1205 10:47:23.871408 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.871384283 +0000 UTC m=+4.874578214 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-certs-ca-bundle" (UniqueName: "kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync configmap cache: timed out waiting for the condition Dec 05 10:47:23.871581 master-0 kubenswrapper[24928]: E1205 10:47:23.871522 24928 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.871697 master-0 kubenswrapper[24928]: E1205 10:47:23.871649 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert podName:b7508cd3-7421-4ab9-be1f-318db2853596 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.871620669 +0000 UTC m=+4.874814570 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert") pod "ingress-canary-knq92" (UID: "b7508cd3-7421-4ab9-be1f-318db2853596") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.872352 master-0 kubenswrapper[24928]: E1205 10:47:23.872315 24928 secret.go:189] Couldn't get secret openshift-monitoring/telemeter-client: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.872459 master-0 kubenswrapper[24928]: E1205 10:47:23.872360 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client podName:384f1c47-f511-4b58-aa7f-71aef6ef91a9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.872350756 +0000 UTC m=+4.875544607 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-telemeter-client" (UniqueName: "kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client") pod "telemeter-client-86cb595668-52qnw" (UID: "384f1c47-f511-4b58-aa7f-71aef6ef91a9") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.872656 master-0 kubenswrapper[24928]: I1205 10:47:23.872628 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-m2gng" Dec 05 10:47:23.872714 master-0 kubenswrapper[24928]: E1205 10:47:23.872681 24928 secret.go:189] Couldn't get secret openshift-monitoring/metrics-server-5ll0c5ruaqfm2: failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.872797 master-0 kubenswrapper[24928]: E1205 10:47:23.872765 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle podName:3aa9a063-322b-4dc6-a724-05a66026160b nodeName:}" failed. No retries permitted until 2025-12-05 10:47:24.872742956 +0000 UTC m=+4.875936877 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "client-ca-bundle" (UniqueName: "kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle") pod "metrics-server-7c46d76dff-z8d8z" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b") : failed to sync secret cache: timed out waiting for the condition Dec 05 10:47:23.894814 master-0 kubenswrapper[24928]: I1205 10:47:23.894739 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-l9vpw" Dec 05 10:47:23.917867 master-0 kubenswrapper[24928]: I1205 10:47:23.916316 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Dec 05 10:47:23.940463 master-0 kubenswrapper[24928]: I1205 10:47:23.939879 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-5ll0c5ruaqfm2" Dec 05 10:47:23.952753 master-0 kubenswrapper[24928]: I1205 10:47:23.952704 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 10:47:23.973559 master-0 kubenswrapper[24928]: I1205 10:47:23.973501 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Dec 05 10:47:23.996390 master-0 kubenswrapper[24928]: I1205 10:47:23.996304 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-scdp6" Dec 05 10:47:24.012849 master-0 kubenswrapper[24928]: I1205 10:47:24.012773 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 10:47:24.034106 master-0 kubenswrapper[24928]: I1205 10:47:24.034053 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 10:47:24.055579 master-0 kubenswrapper[24928]: I1205 10:47:24.054986 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Dec 05 10:47:24.072501 master-0 kubenswrapper[24928]: I1205 10:47:24.072412 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Dec 05 10:47:24.101053 master-0 kubenswrapper[24928]: I1205 10:47:24.100941 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-5972r" Dec 05 10:47:24.113814 master-0 kubenswrapper[24928]: I1205 10:47:24.113098 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 10:47:24.136870 master-0 kubenswrapper[24928]: I1205 10:47:24.134955 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client-tls" Dec 05 10:47:24.153533 master-0 kubenswrapper[24928]: I1205 10:47:24.153453 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client-dockercfg-2kvvs" Dec 05 10:47:24.172885 master-0 kubenswrapper[24928]: I1205 10:47:24.172803 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client" Dec 05 10:47:24.192444 master-0 kubenswrapper[24928]: I1205 10:47:24.192285 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client-kube-rbac-proxy-config" Dec 05 10:47:24.211836 master-0 kubenswrapper[24928]: I1205 10:47:24.210670 24928 request.go:700] Waited for 2.999432652s due to client-side throttling, not priority and fairness, request: GET:https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-monitoring/secrets?fieldSelector=metadata.name%3Dfederate-client-certs&limit=500&resourceVersion=0 Dec 05 10:47:24.214949 master-0 kubenswrapper[24928]: I1205 10:47:24.212135 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"federate-client-certs" Dec 05 10:47:24.233892 master-0 kubenswrapper[24928]: I1205 10:47:24.233821 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemeter-client-serving-certs-ca-bundle" Dec 05 10:47:24.271775 master-0 kubenswrapper[24928]: I1205 10:47:24.270473 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemeter-trusted-ca-bundle-56c9b9fa8d9gs" Dec 05 10:47:24.333447 master-0 kubenswrapper[24928]: I1205 10:47:24.330550 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n7tf\" (UniqueName: \"kubernetes.io/projected/720a7467-ce93-4d48-82ec-9ad0922d99c2-kube-api-access-5n7tf\") pod \"network-check-target-d6fzk\" (UID: \"720a7467-ce93-4d48-82ec-9ad0922d99c2\") " pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:47:24.338452 master-0 kubenswrapper[24928]: I1205 10:47:24.337465 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m88gw\" (UniqueName: \"kubernetes.io/projected/d5198438-06ae-4e63-a7e3-950ba23bba9c-kube-api-access-m88gw\") pod \"apiserver-5b9fd577f8-6sxcx\" (UID: \"d5198438-06ae-4e63-a7e3-950ba23bba9c\") " pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:24.353585 master-0 kubenswrapper[24928]: I1205 10:47:24.353451 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dtfn\" (UniqueName: \"kubernetes.io/projected/5a1bdc70-6412-47e0-8330-04d796cc8d55-kube-api-access-5dtfn\") pod \"dns-operator-7c56cf9b74-x6t9h\" (UID: \"5a1bdc70-6412-47e0-8330-04d796cc8d55\") " pod="openshift-dns-operator/dns-operator-7c56cf9b74-x6t9h" Dec 05 10:47:24.380061 master-0 kubenswrapper[24928]: I1205 10:47:24.380002 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f69hj\" (UniqueName: \"kubernetes.io/projected/48bd1d86-a6f2-439f-ab04-6a9a442bec42-kube-api-access-f69hj\") pod \"cluster-baremetal-operator-78f758c7b9-6t2gm\" (UID: \"48bd1d86-a6f2-439f-ab04-6a9a442bec42\") " pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" Dec 05 10:47:24.390369 master-0 kubenswrapper[24928]: I1205 10:47:24.390315 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/8c649a16-c187-412e-b5da-62a00bee38ab-kube-api-access-d8pv2\") pod \"package-server-manager-67477646d4-nm8cn\" (UID: \"8c649a16-c187-412e-b5da-62a00bee38ab\") " pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:47:24.409170 master-0 kubenswrapper[24928]: I1205 10:47:24.409046 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4jgc\" (UniqueName: \"kubernetes.io/projected/0a4db531-8af7-4085-a6b2-0de51b527ce6-kube-api-access-g4jgc\") pod \"redhat-operators-pqhfn\" (UID: \"0a4db531-8af7-4085-a6b2-0de51b527ce6\") " pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:47:24.424227 master-0 kubenswrapper[24928]: I1205 10:47:24.424180 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8wnp\" (UniqueName: \"kubernetes.io/projected/2c3fab4d-05b8-42ec-8c91-91ff64562649-kube-api-access-l8wnp\") pod \"packageserver-d7b67d8cf-krp6c\" (UID: \"2c3fab4d-05b8-42ec-8c91-91ff64562649\") " pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:24.445113 master-0 kubenswrapper[24928]: I1205 10:47:24.445068 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkqz7\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-kube-api-access-vkqz7\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:47:24.465014 master-0 kubenswrapper[24928]: I1205 10:47:24.464964 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7824l\" (UniqueName: \"kubernetes.io/projected/da9d48c9-6346-4c9f-a690-f7419499c3e6-kube-api-access-7824l\") pod \"network-check-source-85d8db45d4-c2mhw\" (UID: \"da9d48c9-6346-4c9f-a690-f7419499c3e6\") " pod="openshift-network-diagnostics/network-check-source-85d8db45d4-c2mhw" Dec 05 10:47:24.488177 master-0 kubenswrapper[24928]: I1205 10:47:24.488122 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwqkb\" (UniqueName: \"kubernetes.io/projected/1a0235af-2cf2-4ad4-b419-764fb56a0107-kube-api-access-cwqkb\") pod \"network-operator-79767b7ff9-t8j2j\" (UID: \"1a0235af-2cf2-4ad4-b419-764fb56a0107\") " pod="openshift-network-operator/network-operator-79767b7ff9-t8j2j" Dec 05 10:47:24.503520 master-0 kubenswrapper[24928]: I1205 10:47:24.503477 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg9zq\" (UniqueName: \"kubernetes.io/projected/d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e-kube-api-access-wg9zq\") pod \"network-node-identity-ql7j7\" (UID: \"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e\") " pod="openshift-network-node-identity/network-node-identity-ql7j7" Dec 05 10:47:24.519150 master-0 kubenswrapper[24928]: I1205 10:47:24.519096 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:24.519316 master-0 kubenswrapper[24928]: I1205 10:47:24.519264 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:24.522721 master-0 kubenswrapper[24928]: I1205 10:47:24.522678 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:24.527373 master-0 kubenswrapper[24928]: I1205 10:47:24.527332 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6z8n\" (UniqueName: \"kubernetes.io/projected/4c2975ec-e33d-4960-a708-277d41c79b15-kube-api-access-d6z8n\") pod \"catalog-operator-fbc6455c4-mbm77\" (UID: \"4c2975ec-e33d-4960-a708-277d41c79b15\") " pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:47:24.548634 master-0 kubenswrapper[24928]: I1205 10:47:24.548566 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rlts\" (UniqueName: \"kubernetes.io/projected/3aa9a063-322b-4dc6-a724-05a66026160b-kube-api-access-5rlts\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:24.565234 master-0 kubenswrapper[24928]: I1205 10:47:24.565178 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1fd0a349-0f66-4c85-95e4-8bbc96648c05-kube-api-access\") pod \"cluster-version-operator-6d5d5dcc89-27xm6\" (UID: \"1fd0a349-0f66-4c85-95e4-8bbc96648c05\") " pod="openshift-cluster-version/cluster-version-operator-6d5d5dcc89-27xm6" Dec 05 10:47:24.582641 master-0 kubenswrapper[24928]: I1205 10:47:24.582596 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn7rj\" (UniqueName: \"kubernetes.io/projected/eb290494-a456-4f0e-9afc-f20abab1a1bf-kube-api-access-qn7rj\") pod \"cluster-node-tuning-operator-85cff47f46-qwx2p\" (UID: \"eb290494-a456-4f0e-9afc-f20abab1a1bf\") " pod="openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-qwx2p" Dec 05 10:47:24.605899 master-0 kubenswrapper[24928]: I1205 10:47:24.605764 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsrkp\" (UniqueName: \"kubernetes.io/projected/33616deb-ea10-4a38-8681-ab023b526b11-kube-api-access-rsrkp\") pod \"tuned-hvh88\" (UID: \"33616deb-ea10-4a38-8681-ab023b526b11\") " pod="openshift-cluster-node-tuning-operator/tuned-hvh88" Dec 05 10:47:24.626982 master-0 kubenswrapper[24928]: I1205 10:47:24.626924 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-784mb\" (UniqueName: \"kubernetes.io/projected/6728a458-098b-44f9-8c8f-b0d76c5825fa-kube-api-access-784mb\") pod \"machine-config-controller-7c6d64c4cd-blwfs\" (UID: \"6728a458-098b-44f9-8c8f-b0d76c5825fa\") " pod="openshift-machine-config-operator/machine-config-controller-7c6d64c4cd-blwfs" Dec 05 10:47:24.724666 master-0 kubenswrapper[24928]: I1205 10:47:24.724619 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g88bl\" (UniqueName: \"kubernetes.io/projected/b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e-kube-api-access-g88bl\") pod \"prometheus-operator-6c74d9cb9f-r787z\" (UID: \"b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e\") " pod="openshift-monitoring/prometheus-operator-6c74d9cb9f-r787z" Dec 05 10:47:24.725710 master-0 kubenswrapper[24928]: I1205 10:47:24.725683 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv4zs\" (UniqueName: \"kubernetes.io/projected/46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082-kube-api-access-jv4zs\") pod \"olm-operator-7cd7dbb44c-d25sk\" (UID: \"46c5ace3-2d9e-40b3-a9ab-fd0ff8b0b082\") " pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:47:24.726341 master-0 kubenswrapper[24928]: I1205 10:47:24.726302 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgtnt\" (UniqueName: \"kubernetes.io/projected/384f1c47-f511-4b58-aa7f-71aef6ef91a9-kube-api-access-lgtnt\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.727200 master-0 kubenswrapper[24928]: I1205 10:47:24.727151 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ml74w\" (UniqueName: \"kubernetes.io/projected/e27c0798-ec1c-43cd-b81b-f77f2f11ad0f-kube-api-access-ml74w\") pod \"csi-snapshot-controller-6b958b6f94-lgn6v\" (UID: \"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" Dec 05 10:47:24.729932 master-0 kubenswrapper[24928]: I1205 10:47:24.729896 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-564nt\" (UniqueName: \"kubernetes.io/projected/baee05cb-62ad-4eda-8a13-e317eae840e5-kube-api-access-564nt\") pod \"node-resolver-qkccw\" (UID: \"baee05cb-62ad-4eda-8a13-e317eae840e5\") " pod="openshift-dns/node-resolver-qkccw" Dec 05 10:47:24.746704 master-0 kubenswrapper[24928]: I1205 10:47:24.746642 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wfsv\" (UniqueName: \"kubernetes.io/projected/1e69ce9e-4e6f-4015-9ba6-5a7942570190-kube-api-access-7wfsv\") pod \"openshift-controller-manager-operator-6c8676f99d-cwvk5\" (UID: \"1e69ce9e-4e6f-4015-9ba6-5a7942570190\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-cwvk5" Dec 05 10:47:24.769974 master-0 kubenswrapper[24928]: I1205 10:47:24.769882 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vp9b\" (UniqueName: \"kubernetes.io/projected/49051e6e-5a2f-45c8-bad0-374514a91c07-kube-api-access-6vp9b\") pod \"cluster-olm-operator-56fcb6cc5f-m6p27\" (UID: \"49051e6e-5a2f-45c8-bad0-374514a91c07\") " pod="openshift-cluster-olm-operator/cluster-olm-operator-56fcb6cc5f-m6p27" Dec 05 10:47:24.791460 master-0 kubenswrapper[24928]: I1205 10:47:24.791386 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:24.794259 master-0 kubenswrapper[24928]: I1205 10:47:24.794205 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpgqz\" (UniqueName: \"kubernetes.io/projected/d9ff1ae2-8547-42db-9fd0-f4782589ca18-kube-api-access-wpgqz\") pod \"openshift-state-metrics-5974b6b869-9p5mt\" (UID: \"d9ff1ae2-8547-42db-9fd0-f4782589ca18\") " pod="openshift-monitoring/openshift-state-metrics-5974b6b869-9p5mt" Dec 05 10:47:24.795184 master-0 kubenswrapper[24928]: I1205 10:47:24.795150 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:24.806346 master-0 kubenswrapper[24928]: I1205 10:47:24.806304 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdc7g\" (UniqueName: \"kubernetes.io/projected/0de26264-46c2-4d29-97da-25a1681d6a8e-kube-api-access-cdc7g\") pod \"cluster-samples-operator-797cfd8b47-glpx7\" (UID: \"0de26264-46c2-4d29-97da-25a1681d6a8e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-797cfd8b47-glpx7" Dec 05 10:47:24.825028 master-0 kubenswrapper[24928]: I1205 10:47:24.824947 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98qt8\" (UniqueName: \"kubernetes.io/projected/c22d947f-a5b6-4f24-b142-dd201c46293b-kube-api-access-98qt8\") pod \"openshift-config-operator-68758cbcdb-dnpcv\" (UID: \"c22d947f-a5b6-4f24-b142-dd201c46293b\") " pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:47:24.846822 master-0 kubenswrapper[24928]: I1205 10:47:24.846755 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lwgq\" (UniqueName: \"kubernetes.io/projected/a722cda9-29a0-4b7f-8e1d-9a8950ed765a-kube-api-access-2lwgq\") pod \"cluster-monitoring-operator-7ff994598c-kq8qr\" (UID: \"a722cda9-29a0-4b7f-8e1d-9a8950ed765a\") " pod="openshift-monitoring/cluster-monitoring-operator-7ff994598c-kq8qr" Dec 05 10:47:24.867000 master-0 kubenswrapper[24928]: I1205 10:47:24.866876 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvw27\" (UniqueName: \"kubernetes.io/projected/b7e7f216-f9af-41df-a0b0-df2a76b9b72a-kube-api-access-rvw27\") pod \"machine-config-daemon-5n6nw\" (UID: \"b7e7f216-f9af-41df-a0b0-df2a76b9b72a\") " pod="openshift-machine-config-operator/machine-config-daemon-5n6nw" Dec 05 10:47:24.886474 master-0 kubenswrapper[24928]: I1205 10:47:24.886387 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xjx8\" (UniqueName: \"kubernetes.io/projected/01488ac4-313d-48d5-9e86-7b2011b9e91e-kube-api-access-2xjx8\") pod \"machine-config-operator-dc5d7666f-2cf9h\" (UID: \"01488ac4-313d-48d5-9e86-7b2011b9e91e\") " pod="openshift-machine-config-operator/machine-config-operator-dc5d7666f-2cf9h" Dec 05 10:47:24.904033 master-0 kubenswrapper[24928]: I1205 10:47:24.903988 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5p8s\" (UniqueName: \"kubernetes.io/projected/62f174ff-9f3c-4051-a34a-e55a59207171-kube-api-access-r5p8s\") pod \"machine-approver-74d9cbffbc-9jbnk\" (UID: \"62f174ff-9f3c-4051-a34a-e55a59207171\") " pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" Dec 05 10:47:24.918400 master-0 kubenswrapper[24928]: I1205 10:47:24.918342 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:24.918581 master-0 kubenswrapper[24928]: I1205 10:47:24.918409 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.918700 master-0 kubenswrapper[24928]: I1205 10:47:24.918659 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:24.918877 master-0 kubenswrapper[24928]: I1205 10:47:24.918837 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:24.918913 master-0 kubenswrapper[24928]: I1205 10:47:24.918874 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.918948 master-0 kubenswrapper[24928]: I1205 10:47:24.918928 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:24.919016 master-0 kubenswrapper[24928]: I1205 10:47:24.918988 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-serving-certs-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.919054 master-0 kubenswrapper[24928]: I1205 10:47:24.919044 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:47:24.919084 master-0 kubenswrapper[24928]: I1205 10:47:24.919075 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.919114 master-0 kubenswrapper[24928]: I1205 10:47:24.919095 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:47:24.919142 master-0 kubenswrapper[24928]: I1205 10:47:24.919125 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:24.919210 master-0 kubenswrapper[24928]: I1205 10:47:24.919182 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client-kube-rbac-proxy-config\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.919291 master-0 kubenswrapper[24928]: I1205 10:47:24.919266 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:24.919535 master-0 kubenswrapper[24928]: I1205 10:47:24.919512 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3332b604-6f0d-4243-bce1-b3ab76e58c78-webhook-certs\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:47:24.919535 master-0 kubenswrapper[24928]: I1205 10:47:24.919513 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.919607 master-0 kubenswrapper[24928]: I1205 10:47:24.919555 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.919639 master-0 kubenswrapper[24928]: I1205 10:47:24.919631 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:24.919700 master-0 kubenswrapper[24928]: I1205 10:47:24.919655 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-telemeter-client\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-secret-telemeter-client\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.919801 master-0 kubenswrapper[24928]: I1205 10:47:24.919773 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemeter-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-trusted-ca-bundle\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.919801 master-0 kubenswrapper[24928]: I1205 10:47:24.919790 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.919897 master-0 kubenswrapper[24928]: I1205 10:47:24.919874 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:24.919963 master-0 kubenswrapper[24928]: I1205 10:47:24.919920 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:24.920012 master-0 kubenswrapper[24928]: I1205 10:47:24.919919 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"federate-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-federate-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.920051 master-0 kubenswrapper[24928]: I1205 10:47:24.920014 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemeter-client-tls\" (UniqueName: \"kubernetes.io/secret/384f1c47-f511-4b58-aa7f-71aef6ef91a9-telemeter-client-tls\") pod \"telemeter-client-86cb595668-52qnw\" (UID: \"384f1c47-f511-4b58-aa7f-71aef6ef91a9\") " pod="openshift-monitoring/telemeter-client-86cb595668-52qnw" Dec 05 10:47:24.920157 master-0 kubenswrapper[24928]: I1205 10:47:24.920129 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls\") pod \"metrics-server-7c46d76dff-z8d8z\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:24.920294 master-0 kubenswrapper[24928]: I1205 10:47:24.920269 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7508cd3-7421-4ab9-be1f-318db2853596-cert\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:47:24.925077 master-0 kubenswrapper[24928]: I1205 10:47:24.924972 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pvrn\" (UniqueName: \"kubernetes.io/projected/489b6771-aa16-469f-9883-9a9233d3c379-kube-api-access-9pvrn\") pod \"certified-operators-djhk8\" (UID: \"489b6771-aa16-469f-9883-9a9233d3c379\") " pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:47:24.948304 master-0 kubenswrapper[24928]: I1205 10:47:24.948256 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv2tg\" (UniqueName: \"kubernetes.io/projected/54f9f9d4-8c31-4636-8010-22ee11b9b323-kube-api-access-sv2tg\") pod \"kube-state-metrics-5857974f64-xj7pj\" (UID: \"54f9f9d4-8c31-4636-8010-22ee11b9b323\") " pod="openshift-monitoring/kube-state-metrics-5857974f64-xj7pj" Dec 05 10:47:24.970841 master-0 kubenswrapper[24928]: I1205 10:47:24.970789 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn6sc\" (UniqueName: \"kubernetes.io/projected/97192d4d-15d3-4740-82f3-d0d45f9fe7b9-kube-api-access-nn6sc\") pod \"service-ca-77c99c46b8-m7zqs\" (UID: \"97192d4d-15d3-4740-82f3-d0d45f9fe7b9\") " pod="openshift-service-ca/service-ca-77c99c46b8-m7zqs" Dec 05 10:47:24.995893 master-0 kubenswrapper[24928]: I1205 10:47:24.995827 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dk2b\" (UniqueName: \"kubernetes.io/projected/82464e92-4de2-48f0-8772-a489abb16898-kube-api-access-4dk2b\") pod \"machine-api-operator-88d48b57d-x7jfs\" (UID: \"82464e92-4de2-48f0-8772-a489abb16898\") " pod="openshift-machine-api/machine-api-operator-88d48b57d-x7jfs" Dec 05 10:47:25.010088 master-0 kubenswrapper[24928]: I1205 10:47:25.010024 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrwt6\" (UniqueName: \"kubernetes.io/projected/24444474-aa9a-4a0f-8b4d-90f0009e0dc7-kube-api-access-lrwt6\") pod \"control-plane-machine-set-operator-7df95c79b5-qnq6t\" (UID: \"24444474-aa9a-4a0f-8b4d-90f0009e0dc7\") " pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" Dec 05 10:47:25.031156 master-0 kubenswrapper[24928]: I1205 10:47:25.031111 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdsh9\" (UniqueName: \"kubernetes.io/projected/3c4f767d-65f9-434b-8ddd-ceb0b91ab99a-kube-api-access-pdsh9\") pod \"migrator-74b7b57c65-sfvzd\" (UID: \"3c4f767d-65f9-434b-8ddd-ceb0b91ab99a\") " pod="openshift-kube-storage-version-migrator/migrator-74b7b57c65-sfvzd" Dec 05 10:47:25.044729 master-0 kubenswrapper[24928]: I1205 10:47:25.044675 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fbtj\" (UniqueName: \"kubernetes.io/projected/58b95ae4-7f74-4777-8441-0a0ae28199e9-kube-api-access-7fbtj\") pod \"cluster-storage-operator-dcf7fc84b-9rzps\" (UID: \"58b95ae4-7f74-4777-8441-0a0ae28199e9\") " pod="openshift-cluster-storage-operator/cluster-storage-operator-dcf7fc84b-9rzps" Dec 05 10:47:25.056045 master-0 kubenswrapper[24928]: I1205 10:47:25.055988 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:47:25.063226 master-0 kubenswrapper[24928]: I1205 10:47:25.063166 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-68758cbcdb-dnpcv" Dec 05 10:47:25.067272 master-0 kubenswrapper[24928]: I1205 10:47:25.067246 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmhhw\" (UniqueName: \"kubernetes.io/projected/a4fd453c-e667-4bdc-aa9e-3d95ff707200-kube-api-access-kmhhw\") pod \"catalogd-controller-manager-7cc89f4c4c-lth87\" (UID: \"a4fd453c-e667-4bdc-aa9e-3d95ff707200\") " pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:25.084043 master-0 kubenswrapper[24928]: I1205 10:47:25.083972 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flvlq\" (UniqueName: \"kubernetes.io/projected/568e9bed-cbd2-49bd-84af-9e17f95a003f-kube-api-access-flvlq\") pod \"machine-config-server-5t4nn\" (UID: \"568e9bed-cbd2-49bd-84af-9e17f95a003f\") " pod="openshift-machine-config-operator/machine-config-server-5t4nn" Dec 05 10:47:25.112634 master-0 kubenswrapper[24928]: I1205 10:47:25.112342 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twh85\" (UniqueName: \"kubernetes.io/projected/f7a08359-0379-4364-8b0c-ddb58ff605f4-kube-api-access-twh85\") pod \"etcd-operator-5bf4d88c6f-n8t5c\" (UID: \"f7a08359-0379-4364-8b0c-ddb58ff605f4\") " pod="openshift-etcd-operator/etcd-operator-5bf4d88c6f-n8t5c" Dec 05 10:47:25.128207 master-0 kubenswrapper[24928]: I1205 10:47:25.128094 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljdt5\" (UniqueName: \"kubernetes.io/projected/af7078c2-ad4f-415b-ba2f-77f5145c3b3f-kube-api-access-ljdt5\") pod \"cluster-cloud-controller-manager-operator-758cf9d97b-74dgz\" (UID: \"af7078c2-ad4f-415b-ba2f-77f5145c3b3f\") " pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" Dec 05 10:47:25.146791 master-0 kubenswrapper[24928]: I1205 10:47:25.146703 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkmvj\" (UniqueName: \"kubernetes.io/projected/d1c3b7dd-f25e-4983-8a94-084f863fd5b9-kube-api-access-fkmvj\") pod \"service-ca-operator-77758bc754-9lzv4\" (UID: \"d1c3b7dd-f25e-4983-8a94-084f863fd5b9\") " pod="openshift-service-ca-operator/service-ca-operator-77758bc754-9lzv4" Dec 05 10:47:25.165373 master-0 kubenswrapper[24928]: I1205 10:47:25.165318 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfz87\" (UniqueName: \"kubernetes.io/projected/b760849c-8d83-47da-8677-68445c143bef-kube-api-access-jfz87\") pod \"ovnkube-control-plane-5df5548d54-gr5gp\" (UID: \"b760849c-8d83-47da-8677-68445c143bef\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" Dec 05 10:47:25.185343 master-0 kubenswrapper[24928]: I1205 10:47:25.185073 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjrgm\" (UniqueName: \"kubernetes.io/projected/87909f47-f2d7-46f8-a1c8-27336cdcce5d-kube-api-access-rjrgm\") pod \"csi-snapshot-controller-operator-6bc8656fdc-vd94f\" (UID: \"87909f47-f2d7-46f8-a1c8-27336cdcce5d\") " pod="openshift-cluster-storage-operator/csi-snapshot-controller-operator-6bc8656fdc-vd94f" Dec 05 10:47:25.203034 master-0 kubenswrapper[24928]: I1205 10:47:25.202972 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-bound-sa-token\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:47:25.223645 master-0 kubenswrapper[24928]: I1205 10:47:25.223576 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjgc4\" (UniqueName: \"kubernetes.io/projected/495ba1ea-f844-43ec-8be7-47e738f5428a-kube-api-access-tjgc4\") pod \"ovnkube-node-rsfjs\" (UID: \"495ba1ea-f844-43ec-8be7-47e738f5428a\") " pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:25.230826 master-0 kubenswrapper[24928]: I1205 10:47:25.230772 24928 request.go:700] Waited for 3.933463997s due to client-side throttling, not priority and fairness, request: POST:https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-multus/serviceaccounts/metrics-daemon-sa/token Dec 05 10:47:25.247084 master-0 kubenswrapper[24928]: I1205 10:47:25.247018 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssssf\" (UniqueName: \"kubernetes.io/projected/3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c-kube-api-access-ssssf\") pod \"network-metrics-daemon-8gjgm\" (UID: \"3a1e2bf3-2e33-4a2a-b306-6d5e1d29727c\") " pod="openshift-multus/network-metrics-daemon-8gjgm" Dec 05 10:47:25.264246 master-0 kubenswrapper[24928]: I1205 10:47:25.264177 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hskm2\" (UniqueName: \"kubernetes.io/projected/445d75af-d072-4fa0-91a7-f3fa579b9ca9-kube-api-access-hskm2\") pod \"router-default-5465c8b4db-s4c2f\" (UID: \"445d75af-d072-4fa0-91a7-f3fa579b9ca9\") " pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:25.287759 master-0 kubenswrapper[24928]: I1205 10:47:25.287698 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvnxf\" (UniqueName: \"kubernetes.io/projected/6a9f011f-36f1-4308-a365-69425c186c7f-kube-api-access-mvnxf\") pod \"multus-lxmgz\" (UID: \"6a9f011f-36f1-4308-a365-69425c186c7f\") " pod="openshift-multus/multus-lxmgz" Dec 05 10:47:25.305820 master-0 kubenswrapper[24928]: I1205 10:47:25.305762 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltjt6\" (UniqueName: \"kubernetes.io/projected/ca45c52e-fb30-4e7c-8c3f-e685c0909916-kube-api-access-ltjt6\") pod \"apiserver-85b8f855df-8g52w\" (UID: \"ca45c52e-fb30-4e7c-8c3f-e685c0909916\") " pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:25.329124 master-0 kubenswrapper[24928]: I1205 10:47:25.329050 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4lbg\" (UniqueName: \"kubernetes.io/projected/88cccb5b-1ad1-4fab-b34c-90252794ee20-kube-api-access-f4lbg\") pod \"route-controller-manager-c7946c9c4-hq97s\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:47:25.348670 master-0 kubenswrapper[24928]: I1205 10:47:25.348611 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22676fac-b770-4937-9bee-7478bd1babb7-bound-sa-token\") pod \"ingress-operator-8649c48786-cgt5x\" (UID: \"22676fac-b770-4937-9bee-7478bd1babb7\") " pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" Dec 05 10:47:25.358740 master-0 kubenswrapper[24928]: I1205 10:47:25.358697 24928 scope.go:117] "RemoveContainer" containerID="76b36235b0c1dc04c6934f55253404fa8e8aa924bd0ff655fd03d44b54d22caa" Dec 05 10:47:25.364695 master-0 kubenswrapper[24928]: I1205 10:47:25.364659 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:25.365605 master-0 kubenswrapper[24928]: I1205 10:47:25.365562 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fbss\" (UniqueName: \"kubernetes.io/projected/d95a56ba-c940-4e3e-aed6-d8c04f1871b6-kube-api-access-2fbss\") pod \"authentication-operator-6c968fdfdf-t7sl8\" (UID: \"d95a56ba-c940-4e3e-aed6-d8c04f1871b6\") " pod="openshift-authentication-operator/authentication-operator-6c968fdfdf-t7sl8" Dec 05 10:47:25.372115 master-0 kubenswrapper[24928]: I1205 10:47:25.369221 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:25.372115 master-0 kubenswrapper[24928]: I1205 10:47:25.369313 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:25.386750 master-0 kubenswrapper[24928]: I1205 10:47:25.386651 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28qvt\" (UniqueName: \"kubernetes.io/projected/baa812e4-b011-41ee-82ab-8f571f0b7e0a-kube-api-access-28qvt\") pod \"community-operators-6p8cq\" (UID: \"baa812e4-b011-41ee-82ab-8f571f0b7e0a\") " pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:47:25.407702 master-0 kubenswrapper[24928]: I1205 10:47:25.407653 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx7jr\" (UniqueName: \"kubernetes.io/projected/58baad85-de54-49e7-a13e-d470d9c50d11-kube-api-access-jx7jr\") pod \"dns-default-4vxng\" (UID: \"58baad85-de54-49e7-a13e-d470d9c50d11\") " pod="openshift-dns/dns-default-4vxng" Dec 05 10:47:25.423248 master-0 kubenswrapper[24928]: I1205 10:47:25.423194 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd87p\" (UniqueName: \"kubernetes.io/projected/592373ae-a952-4dd3-a9bc-f9c9c19c0802-kube-api-access-rd87p\") pod \"operator-controller-controller-manager-7cbd59c7f8-dh5tt\" (UID: \"592373ae-a952-4dd3-a9bc-f9c9c19c0802\") " pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:25.444414 master-0 kubenswrapper[24928]: I1205 10:47:25.444340 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrdsv\" (UniqueName: \"kubernetes.io/projected/38ad6e6a-a2b9-44e9-ac0b-e413c65efad8-kube-api-access-jrdsv\") pod \"multus-additional-cni-plugins-dms5d\" (UID: \"38ad6e6a-a2b9-44e9-ac0b-e413c65efad8\") " pod="openshift-multus/multus-additional-cni-plugins-dms5d" Dec 05 10:47:25.463508 master-0 kubenswrapper[24928]: I1205 10:47:25.463462 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkzkh\" (UniqueName: \"kubernetes.io/projected/1f132702-304b-46ff-b428-6df1deeffec3-kube-api-access-vkzkh\") pod \"cluster-autoscaler-operator-5f49d774cd-cfg5f\" (UID: \"1f132702-304b-46ff-b428-6df1deeffec3\") " pod="openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-cfg5f" Dec 05 10:47:25.489194 master-0 kubenswrapper[24928]: I1205 10:47:25.489073 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:25.489473 master-0 kubenswrapper[24928]: I1205 10:47:25.489244 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:25.491084 master-0 kubenswrapper[24928]: I1205 10:47:25.491038 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ktnb\" (UniqueName: \"kubernetes.io/projected/ee0b4a7f-1b96-4304-bef0-fb575e0e431c-kube-api-access-9ktnb\") pod \"cloud-credential-operator-698c598cfc-rgc4p\" (UID: \"ee0b4a7f-1b96-4304-bef0-fb575e0e431c\") " pod="openshift-cloud-credential-operator/cloud-credential-operator-698c598cfc-rgc4p" Dec 05 10:47:25.505245 master-0 kubenswrapper[24928]: I1205 10:47:25.505199 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8bpw\" (UniqueName: \"kubernetes.io/projected/21a9bd5a-3ade-48ef-8004-e0492daa85ba-kube-api-access-r8bpw\") pod \"node-exporter-bmqsb\" (UID: \"21a9bd5a-3ade-48ef-8004-e0492daa85ba\") " pod="openshift-monitoring/node-exporter-bmqsb" Dec 05 10:47:25.525512 master-0 kubenswrapper[24928]: I1205 10:47:25.525410 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fncz7\" (UniqueName: \"kubernetes.io/projected/926263c4-ec5b-41cb-9c30-0c88f636035f-kube-api-access-fncz7\") pod \"marketplace-operator-f797b99b6-z9qcl\" (UID: \"926263c4-ec5b-41cb-9c30-0c88f636035f\") " pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:47:25.543326 master-0 kubenswrapper[24928]: I1205 10:47:25.543277 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/11f563d5-89bb-433c-956a-6d5d2492e8f1-kube-api-access\") pod \"kube-controller-manager-operator-848f645654-rmdb8\" (UID: \"11f563d5-89bb-433c-956a-6d5d2492e8f1\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-rmdb8" Dec 05 10:47:25.563213 master-0 kubenswrapper[24928]: I1205 10:47:25.563166 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnzgs\" (UniqueName: \"kubernetes.io/projected/0aa6e138-4b1f-4047-8255-a2b14d044588-kube-api-access-wnzgs\") pod \"redhat-marketplace-wk29h\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:47:25.582440 master-0 kubenswrapper[24928]: I1205 10:47:25.582361 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gfgm\" (UniqueName: \"kubernetes.io/projected/3332b604-6f0d-4243-bce1-b3ab76e58c78-kube-api-access-6gfgm\") pod \"multus-admission-controller-8dbbb5754-7p9c2\" (UID: \"3332b604-6f0d-4243-bce1-b3ab76e58c78\") " pod="openshift-multus/multus-admission-controller-8dbbb5754-7p9c2" Dec 05 10:47:25.603695 master-0 kubenswrapper[24928]: I1205 10:47:25.603648 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtsw9\" (UniqueName: \"kubernetes.io/projected/65e06ca6-1f15-43ec-b8dc-a300e1cf83bd-kube-api-access-vtsw9\") pod \"insights-operator-55965856b6-2sxv7\" (UID: \"65e06ca6-1f15-43ec-b8dc-a300e1cf83bd\") " pod="openshift-insights/insights-operator-55965856b6-2sxv7" Dec 05 10:47:25.626553 master-0 kubenswrapper[24928]: I1205 10:47:25.626458 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2vd4\" (UniqueName: \"kubernetes.io/projected/fd58232c-a81a-4aee-8b2c-5ffcdded2e23-kube-api-access-h2vd4\") pod \"kube-storage-version-migrator-operator-b9c5dfc78-4gqxr\" (UID: \"fd58232c-a81a-4aee-8b2c-5ffcdded2e23\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b9c5dfc78-4gqxr" Dec 05 10:47:25.646511 master-0 kubenswrapper[24928]: I1205 10:47:25.646374 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khfxz\" (UniqueName: \"kubernetes.io/projected/f7b29f89-e42d-4e53-ad14-05efdce933f0-kube-api-access-khfxz\") pod \"cluster-image-registry-operator-6fb9f88b7-f29mb\" (UID: \"f7b29f89-e42d-4e53-ad14-05efdce933f0\") " pod="openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-f29mb" Dec 05 10:47:25.663688 master-0 kubenswrapper[24928]: I1205 10:47:25.663590 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-4vxng" Dec 05 10:47:25.663957 master-0 kubenswrapper[24928]: I1205 10:47:25.663897 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95jmp\" (UniqueName: \"kubernetes.io/projected/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-kube-api-access-95jmp\") pod \"controller-manager-86f4478dbf-jqlt9\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:25.664665 master-0 kubenswrapper[24928]: I1205 10:47:25.664615 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-4vxng" Dec 05 10:47:25.683868 master-0 kubenswrapper[24928]: I1205 10:47:25.683787 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4825316a-ea9f-4d3d-838b-fa809a6e49c7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5f85974995-dwh5t\" (UID: \"4825316a-ea9f-4d3d-838b-fa809a6e49c7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5f85974995-dwh5t" Dec 05 10:47:25.704899 master-0 kubenswrapper[24928]: I1205 10:47:25.704844 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgvv6\" (UniqueName: \"kubernetes.io/projected/b7508cd3-7421-4ab9-be1f-318db2853596-kube-api-access-kgvv6\") pod \"ingress-canary-knq92\" (UID: \"b7508cd3-7421-4ab9-be1f-318db2853596\") " pod="openshift-ingress-canary/ingress-canary-knq92" Dec 05 10:47:25.725726 master-0 kubenswrapper[24928]: I1205 10:47:25.725675 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzwfq\" (UniqueName: \"kubernetes.io/projected/6f76d12f-5406-47e2-8337-2f50e35376d6-kube-api-access-fzwfq\") pod \"openshift-apiserver-operator-7bf7f6b755-hdjv7\" (UID: \"6f76d12f-5406-47e2-8337-2f50e35376d6\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-hdjv7" Dec 05 10:47:25.744076 master-0 kubenswrapper[24928]: I1205 10:47:25.744024 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhjhk\" (UniqueName: \"kubernetes.io/projected/283122ba-be1c-4516-bd0f-df41c13c098b-kube-api-access-vhjhk\") pod \"iptables-alerter-d6wjk\" (UID: \"283122ba-be1c-4516-bd0f-df41c13c098b\") " pod="openshift-network-operator/iptables-alerter-d6wjk" Dec 05 10:47:25.763204 master-0 kubenswrapper[24928]: I1205 10:47:25.763149 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/444f8808-e454-4015-9e20-429e715a08c7-kube-api-access\") pod \"kube-apiserver-operator-765d9ff747-p57fl\" (UID: \"444f8808-e454-4015-9e20-429e715a08c7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-p57fl" Dec 05 10:47:25.788021 master-0 kubenswrapper[24928]: E1205 10:47:25.787962 24928 projected.go:288] Couldn't get configMap openshift-etcd/kube-root-ca.crt: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:25.788021 master-0 kubenswrapper[24928]: E1205 10:47:25.788012 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-etcd/installer-2-master-0: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:25.788312 master-0 kubenswrapper[24928]: E1205 10:47:25.788107 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access podName:68058d25-5888-495f-8571-453f0bf919d9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:26.288083104 +0000 UTC m=+6.291276955 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access") pod "installer-2-master-0" (UID: "68058d25-5888-495f-8571-453f0bf919d9") : object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:25.789558 master-0 kubenswrapper[24928]: I1205 10:47:25.789496 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:25.803702 master-0 kubenswrapper[24928]: E1205 10:47:25.803632 24928 projected.go:288] Couldn't get configMap openshift-kube-apiserver/kube-root-ca.crt: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:25.803702 master-0 kubenswrapper[24928]: E1205 10:47:25.803675 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-apiserver/installer-3-master-0: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:25.803920 master-0 kubenswrapper[24928]: E1205 10:47:25.803737 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access podName:75ba20a1-b322-4ce8-88d2-d8ee210cc27a nodeName:}" failed. No retries permitted until 2025-12-05 10:47:26.303719094 +0000 UTC m=+6.306912945 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access") pod "installer-3-master-0" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a") : object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:25.853719 master-0 kubenswrapper[24928]: I1205 10:47:25.853665 24928 kubelet_node_status.go:115] "Node was previously registered" node="master-0" Dec 05 10:47:25.853916 master-0 kubenswrapper[24928]: I1205 10:47:25.853793 24928 kubelet_node_status.go:79] "Successfully registered node" node="master-0" Dec 05 10:47:25.895227 master-0 kubenswrapper[24928]: I1205 10:47:25.895174 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" Dec 05 10:47:25.899809 master-0 kubenswrapper[24928]: I1205 10:47:25.899724 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-vjvbz" Dec 05 10:47:25.922511 master-0 kubenswrapper[24928]: E1205 10:47:25.922455 24928 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-rbac-proxy-crio-master-0\" already exists" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-master-0" Dec 05 10:47:26.344623 master-0 kubenswrapper[24928]: I1205 10:47:26.344546 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:26.344623 master-0 kubenswrapper[24928]: I1205 10:47:26.344634 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:26.345188 master-0 kubenswrapper[24928]: E1205 10:47:26.344832 24928 projected.go:288] Couldn't get configMap openshift-etcd/kube-root-ca.crt: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:26.345188 master-0 kubenswrapper[24928]: E1205 10:47:26.344879 24928 projected.go:288] Couldn't get configMap openshift-kube-apiserver/kube-root-ca.crt: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:26.345188 master-0 kubenswrapper[24928]: E1205 10:47:26.344926 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-apiserver/installer-3-master-0: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:26.345188 master-0 kubenswrapper[24928]: E1205 10:47:26.344985 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access podName:75ba20a1-b322-4ce8-88d2-d8ee210cc27a nodeName:}" failed. No retries permitted until 2025-12-05 10:47:27.344962977 +0000 UTC m=+7.348156828 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access") pod "installer-3-master-0" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a") : object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:26.345188 master-0 kubenswrapper[24928]: E1205 10:47:26.344889 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-etcd/installer-2-master-0: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:26.345188 master-0 kubenswrapper[24928]: E1205 10:47:26.345116 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access podName:68058d25-5888-495f-8571-453f0bf919d9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:27.345083089 +0000 UTC m=+7.348276970 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access") pod "installer-2-master-0" (UID: "68058d25-5888-495f-8571-453f0bf919d9") : object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:26.367777 master-0 kubenswrapper[24928]: I1205 10:47:26.367716 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:26.495521 master-0 kubenswrapper[24928]: I1205 10:47:26.495473 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/3.log" Dec 05 10:47:26.496130 master-0 kubenswrapper[24928]: I1205 10:47:26.496087 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:26.496186 master-0 kubenswrapper[24928]: I1205 10:47:26.496158 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-8649c48786-cgt5x" event={"ID":"22676fac-b770-4937-9bee-7478bd1babb7","Type":"ContainerStarted","Data":"2852a3d978e502bd201233581f5bfd5a9bee75c0824699a98cf6ea4bf114355d"} Dec 05 10:47:27.065172 master-0 kubenswrapper[24928]: I1205 10:47:27.064851 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" podStartSLOduration=7.064833867 podStartE2EDuration="7.064833867s" podCreationTimestamp="2025-12-05 10:47:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:47:27.064498429 +0000 UTC m=+7.067692280" watchObservedRunningTime="2025-12-05 10:47:27.064833867 +0000 UTC m=+7.068027718" Dec 05 10:47:27.212316 master-0 kubenswrapper[24928]: I1205 10:47:27.212194 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-master-0" podStartSLOduration=8.212163843999999 podStartE2EDuration="8.212163844s" podCreationTimestamp="2025-12-05 10:47:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:47:27.211561489 +0000 UTC m=+7.214755380" watchObservedRunningTime="2025-12-05 10:47:27.212163844 +0000 UTC m=+7.215357735" Dec 05 10:47:27.314198 master-0 kubenswrapper[24928]: I1205 10:47:27.314132 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:27.320310 master-0 kubenswrapper[24928]: I1205 10:47:27.320195 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:27.359958 master-0 kubenswrapper[24928]: I1205 10:47:27.359903 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:27.360489 master-0 kubenswrapper[24928]: E1205 10:47:27.360071 24928 projected.go:288] Couldn't get configMap openshift-etcd/kube-root-ca.crt: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:27.360489 master-0 kubenswrapper[24928]: E1205 10:47:27.360094 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-etcd/installer-2-master-0: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:27.360489 master-0 kubenswrapper[24928]: E1205 10:47:27.360172 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access podName:68058d25-5888-495f-8571-453f0bf919d9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:29.360153575 +0000 UTC m=+9.363347426 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access") pod "installer-2-master-0" (UID: "68058d25-5888-495f-8571-453f0bf919d9") : object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:27.360489 master-0 kubenswrapper[24928]: I1205 10:47:27.360252 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:27.360779 master-0 kubenswrapper[24928]: E1205 10:47:27.360740 24928 projected.go:288] Couldn't get configMap openshift-kube-apiserver/kube-root-ca.crt: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:27.360779 master-0 kubenswrapper[24928]: E1205 10:47:27.360771 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-apiserver/installer-3-master-0: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:27.360890 master-0 kubenswrapper[24928]: E1205 10:47:27.360809 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access podName:75ba20a1-b322-4ce8-88d2-d8ee210cc27a nodeName:}" failed. No retries permitted until 2025-12-05 10:47:29.360796941 +0000 UTC m=+9.363990902 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access") pod "installer-3-master-0" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a") : object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:27.600018 master-0 kubenswrapper[24928]: I1205 10:47:27.599848 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:47:28.049844 master-0 kubenswrapper[24928]: I1205 10:47:28.049770 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:47:28.107763 master-0 kubenswrapper[24928]: I1205 10:47:28.107685 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:47:28.466946 master-0 kubenswrapper[24928]: I1205 10:47:28.466816 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:28.467879 master-0 kubenswrapper[24928]: I1205 10:47:28.467852 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:47:28.589577 master-0 kubenswrapper[24928]: I1205 10:47:28.589494 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:28.596681 master-0 kubenswrapper[24928]: I1205 10:47:28.596590 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-5b9fd577f8-6sxcx" Dec 05 10:47:28.745396 master-0 kubenswrapper[24928]: I1205 10:47:28.745217 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:28.781442 master-0 kubenswrapper[24928]: I1205 10:47:28.781362 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:29.388901 master-0 kubenswrapper[24928]: I1205 10:47:29.388809 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:29.389218 master-0 kubenswrapper[24928]: E1205 10:47:29.389033 24928 projected.go:288] Couldn't get configMap openshift-kube-apiserver/kube-root-ca.crt: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:29.389218 master-0 kubenswrapper[24928]: E1205 10:47:29.389083 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-apiserver/installer-3-master-0: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:29.389218 master-0 kubenswrapper[24928]: I1205 10:47:29.389040 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:29.389411 master-0 kubenswrapper[24928]: E1205 10:47:29.389158 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access podName:75ba20a1-b322-4ce8-88d2-d8ee210cc27a nodeName:}" failed. No retries permitted until 2025-12-05 10:47:33.389131378 +0000 UTC m=+13.392325269 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access") pod "installer-3-master-0" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a") : object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:29.389411 master-0 kubenswrapper[24928]: E1205 10:47:29.389252 24928 projected.go:288] Couldn't get configMap openshift-etcd/kube-root-ca.crt: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:29.389411 master-0 kubenswrapper[24928]: E1205 10:47:29.389350 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-etcd/installer-2-master-0: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:29.389656 master-0 kubenswrapper[24928]: E1205 10:47:29.389493 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access podName:68058d25-5888-495f-8571-453f0bf919d9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:33.389464177 +0000 UTC m=+13.392658038 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access") pod "installer-2-master-0" (UID: "68058d25-5888-495f-8571-453f0bf919d9") : object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:29.518629 master-0 kubenswrapper[24928]: I1205 10:47:29.518543 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:29.518629 master-0 kubenswrapper[24928]: I1205 10:47:29.518602 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:29.576480 master-0 kubenswrapper[24928]: I1205 10:47:29.576405 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:29.580654 master-0 kubenswrapper[24928]: I1205 10:47:29.580585 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:29.580867 master-0 kubenswrapper[24928]: I1205 10:47:29.580844 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:29.597631 master-0 kubenswrapper[24928]: I1205 10:47:29.597554 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:29.639033 master-0 kubenswrapper[24928]: I1205 10:47:29.638926 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:29.708436 master-0 kubenswrapper[24928]: I1205 10:47:29.708341 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:29.711685 master-0 kubenswrapper[24928]: I1205 10:47:29.711631 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:47:29.779612 master-0 kubenswrapper[24928]: I1205 10:47:29.779552 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:29.780049 master-0 kubenswrapper[24928]: I1205 10:47:29.779988 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:29.787331 master-0 kubenswrapper[24928]: I1205 10:47:29.787284 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:47:30.377923 master-0 kubenswrapper[24928]: I1205 10:47:30.377848 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:30.385340 master-0 kubenswrapper[24928]: I1205 10:47:30.385278 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-85b8f855df-8g52w" Dec 05 10:47:30.532504 master-0 kubenswrapper[24928]: I1205 10:47:30.526596 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:30.549457 master-0 kubenswrapper[24928]: I1205 10:47:30.543550 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:30.549457 master-0 kubenswrapper[24928]: I1205 10:47:30.549374 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:47:30.941043 master-0 kubenswrapper[24928]: I1205 10:47:30.940971 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:30.944750 master-0 kubenswrapper[24928]: I1205 10:47:30.944645 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d7b67d8cf-krp6c" Dec 05 10:47:31.040154 master-0 kubenswrapper[24928]: I1205 10:47:31.040095 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-master-0" Dec 05 10:47:31.050440 master-0 kubenswrapper[24928]: I1205 10:47:31.050394 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-master-0" Dec 05 10:47:31.095856 master-0 kubenswrapper[24928]: I1205 10:47:31.095792 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:47:31.100386 master-0 kubenswrapper[24928]: I1205 10:47:31.100336 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:47:31.298000 master-0 kubenswrapper[24928]: I1205 10:47:31.297939 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:47:31.311576 master-0 kubenswrapper[24928]: I1205 10:47:31.311515 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:31.343704 master-0 kubenswrapper[24928]: I1205 10:47:31.343654 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:47:31.372713 master-0 kubenswrapper[24928]: I1205 10:47:31.372665 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:47:31.395259 master-0 kubenswrapper[24928]: I1205 10:47:31.392375 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-djhk8" Dec 05 10:47:31.531285 master-0 kubenswrapper[24928]: I1205 10:47:31.531221 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:31.538875 master-0 kubenswrapper[24928]: I1205 10:47:31.536276 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:47:31.548137 master-0 kubenswrapper[24928]: I1205 10:47:31.546781 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-master-0" Dec 05 10:47:31.577976 master-0 kubenswrapper[24928]: I1205 10:47:31.577920 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:47:31.581259 master-0 kubenswrapper[24928]: I1205 10:47:31.581215 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-67477646d4-nm8cn" Dec 05 10:47:31.956042 master-0 kubenswrapper[24928]: I1205 10:47:31.955912 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:47:32.004672 master-0 kubenswrapper[24928]: I1205 10:47:32.004601 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pqhfn" Dec 05 10:47:32.044235 master-0 kubenswrapper[24928]: I1205 10:47:32.044171 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:47:32.047672 master-0 kubenswrapper[24928]: I1205 10:47:32.047647 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-d25sk" Dec 05 10:47:32.255303 master-0 kubenswrapper[24928]: I1205 10:47:32.255184 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:47:32.260023 master-0 kubenswrapper[24928]: I1205 10:47:32.259992 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-d6fzk" Dec 05 10:47:32.668185 master-0 kubenswrapper[24928]: I1205 10:47:32.668139 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:33.351023 master-0 kubenswrapper[24928]: I1205 10:47:33.350950 24928 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0"] Dec 05 10:47:33.351400 master-0 kubenswrapper[24928]: I1205 10:47:33.351352 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" podUID="a906debd0c35952850935aee2d607cce" containerName="startup-monitor" containerID="cri-o://8930f7af011c1e43a64ba6a4ca85659538a4774816c777a7c34e7a4b10c3cab3" gracePeriod=5 Dec 05 10:47:33.452217 master-0 kubenswrapper[24928]: I1205 10:47:33.452156 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:47:33.454238 master-0 kubenswrapper[24928]: I1205 10:47:33.454165 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:33.454324 master-0 kubenswrapper[24928]: I1205 10:47:33.454291 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:33.454407 master-0 kubenswrapper[24928]: E1205 10:47:33.454377 24928 projected.go:288] Couldn't get configMap openshift-kube-apiserver/kube-root-ca.crt: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:33.454468 master-0 kubenswrapper[24928]: E1205 10:47:33.454412 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-apiserver/installer-3-master-0: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:33.454514 master-0 kubenswrapper[24928]: E1205 10:47:33.454484 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access podName:75ba20a1-b322-4ce8-88d2-d8ee210cc27a nodeName:}" failed. No retries permitted until 2025-12-05 10:47:41.454469654 +0000 UTC m=+21.457663505 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access") pod "installer-3-master-0" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a") : object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:33.454571 master-0 kubenswrapper[24928]: E1205 10:47:33.454377 24928 projected.go:288] Couldn't get configMap openshift-etcd/kube-root-ca.crt: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:33.454571 master-0 kubenswrapper[24928]: E1205 10:47:33.454547 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-etcd/installer-2-master-0: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:33.454650 master-0 kubenswrapper[24928]: E1205 10:47:33.454580 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access podName:68058d25-5888-495f-8571-453f0bf919d9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:41.454569007 +0000 UTC m=+21.457762858 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access") pod "installer-2-master-0" (UID: "68058d25-5888-495f-8571-453f0bf919d9") : object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:33.456458 master-0 kubenswrapper[24928]: I1205 10:47:33.456398 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-mbm77" Dec 05 10:47:33.457772 master-0 kubenswrapper[24928]: I1205 10:47:33.457735 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:47:33.460404 master-0 kubenswrapper[24928]: I1205 10:47:33.460374 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:47:34.210954 master-0 kubenswrapper[24928]: I1205 10:47:34.210915 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:34.211615 master-0 kubenswrapper[24928]: I1205 10:47:34.211596 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:34.215469 master-0 kubenswrapper[24928]: I1205 10:47:34.215381 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5465c8b4db-s4c2f" Dec 05 10:47:34.377175 master-0 kubenswrapper[24928]: I1205 10:47:34.376445 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:47:34.733447 master-0 kubenswrapper[24928]: I1205 10:47:34.733383 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:47:34.785038 master-0 kubenswrapper[24928]: I1205 10:47:34.784963 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:47:34.830600 master-0 kubenswrapper[24928]: I1205 10:47:34.830513 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:47:34.872634 master-0 kubenswrapper[24928]: I1205 10:47:34.872594 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 10:47:36.468688 master-0 kubenswrapper[24928]: I1205 10:47:36.468577 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-77b5b8969c-5clks"] Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.468838 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c5922c-65b0-4c75-992c-f9c88f81762d" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.468852 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c5922c-65b0-4c75-992c-f9c88f81762d" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.468864 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="738d9fe7-2d08-4aa4-842b-9cbeddf120fb" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.468871 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="738d9fe7-2d08-4aa4-842b-9cbeddf120fb" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.468880 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70fbf150-c1f7-4857-8b63-c566282e2526" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.468889 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="70fbf150-c1f7-4857-8b63-c566282e2526" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.468905 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="cluster-policy-controller" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.468917 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="cluster-policy-controller" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.468930 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e149cbe7-1ec3-4a06-af61-3a8906b8e9ef" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.468939 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e149cbe7-1ec3-4a06-af61-3a8906b8e9ef" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.468948 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b60e9f1-b49e-4804-b56d-857c7bb2a18b" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.468955 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b60e9f1-b49e-4804-b56d-857c7bb2a18b" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.468963 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-recovery-controller" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.468969 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-recovery-controller" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.468978 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1d74f88-1419-431e-80da-26db419f050e" containerName="assisted-installer-controller" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.468984 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1d74f88-1419-431e-80da-26db419f050e" containerName="assisted-installer-controller" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.468992 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7db4c690-9d13-4c82-9565-f974d554916b" containerName="collect-profiles" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.468999 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="7db4c690-9d13-4c82-9565-f974d554916b" containerName="collect-profiles" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.469007 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469013 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.469024 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68058d25-5888-495f-8571-453f0bf919d9" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469030 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="68058d25-5888-495f-8571-453f0bf919d9" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.469039 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d8c854f-31f8-42d3-b28c-82c7d1b14ee9" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469045 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d8c854f-31f8-42d3-b28c-82c7d1b14ee9" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.469052 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a906debd0c35952850935aee2d607cce" containerName="startup-monitor" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469058 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="a906debd0c35952850935aee2d607cce" containerName="startup-monitor" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.469073 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75ba20a1-b322-4ce8-88d2-d8ee210cc27a" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469079 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="75ba20a1-b322-4ce8-88d2-d8ee210cc27a" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.469088 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6325f9a2-86d5-4fac-9e71-331b2628f965" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469094 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6325f9a2-86d5-4fac-9e71-331b2628f965" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: E1205 10:47:36.469108 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-cert-syncer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469115 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-cert-syncer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469245 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="70fbf150-c1f7-4857-8b63-c566282e2526" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469261 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="738d9fe7-2d08-4aa4-842b-9cbeddf120fb" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469269 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="68058d25-5888-495f-8571-453f0bf919d9" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469278 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="cluster-policy-controller" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469288 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="a906debd0c35952850935aee2d607cce" containerName="startup-monitor" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469301 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-cert-syncer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469313 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3c5922c-65b0-4c75-992c-f9c88f81762d" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469326 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="7db4c690-9d13-4c82-9565-f974d554916b" containerName="collect-profiles" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469338 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469349 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d8c854f-31f8-42d3-b28c-82c7d1b14ee9" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469359 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1d74f88-1419-431e-80da-26db419f050e" containerName="assisted-installer-controller" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469371 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="6325f9a2-86d5-4fac-9e71-331b2628f965" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469382 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="e149cbe7-1ec3-4a06-af61-3a8906b8e9ef" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469391 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="a51fa80258077cf0f6fddf343b4ba54f" containerName="kube-controller-manager-recovery-controller" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469399 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="75ba20a1-b322-4ce8-88d2-d8ee210cc27a" containerName="installer" Dec 05 10:47:36.469557 master-0 kubenswrapper[24928]: I1205 10:47:36.469407 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b60e9f1-b49e-4804-b56d-857c7bb2a18b" containerName="installer" Dec 05 10:47:36.471545 master-0 kubenswrapper[24928]: I1205 10:47:36.469825 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.473233 master-0 kubenswrapper[24928]: I1205 10:47:36.473197 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 10:47:36.474028 master-0 kubenswrapper[24928]: I1205 10:47:36.473999 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 10:47:36.474157 master-0 kubenswrapper[24928]: I1205 10:47:36.474117 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 10:47:36.474258 master-0 kubenswrapper[24928]: I1205 10:47:36.474119 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 10:47:36.474305 master-0 kubenswrapper[24928]: I1205 10:47:36.474232 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-44p5j" Dec 05 10:47:36.474410 master-0 kubenswrapper[24928]: I1205 10:47:36.474393 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 10:47:36.474547 master-0 kubenswrapper[24928]: I1205 10:47:36.474516 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 10:47:36.474617 master-0 kubenswrapper[24928]: I1205 10:47:36.474556 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 10:47:36.474809 master-0 kubenswrapper[24928]: I1205 10:47:36.474786 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 10:47:36.475077 master-0 kubenswrapper[24928]: I1205 10:47:36.475047 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 10:47:36.475193 master-0 kubenswrapper[24928]: I1205 10:47:36.475172 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 10:47:36.475257 master-0 kubenswrapper[24928]: I1205 10:47:36.475177 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 10:47:36.480845 master-0 kubenswrapper[24928]: I1205 10:47:36.480785 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-77b5b8969c-5clks"] Dec 05 10:47:36.485882 master-0 kubenswrapper[24928]: I1205 10:47:36.485819 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 10:47:36.492901 master-0 kubenswrapper[24928]: I1205 10:47:36.492854 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497013 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-policies\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497075 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-login\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497109 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-error\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497130 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-router-certs\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497172 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-serving-cert\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497198 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497218 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497238 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24l8d\" (UniqueName: \"kubernetes.io/projected/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-kube-api-access-24l8d\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497266 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-dir\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497286 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-service-ca\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497306 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497321 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.497949 master-0 kubenswrapper[24928]: I1205 10:47:36.497341 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.598339 master-0 kubenswrapper[24928]: I1205 10:47:36.598113 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-error\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.598339 master-0 kubenswrapper[24928]: I1205 10:47:36.598216 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-router-certs\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.598737 master-0 kubenswrapper[24928]: I1205 10:47:36.598700 24928 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 05 10:47:36.599046 master-0 kubenswrapper[24928]: I1205 10:47:36.598712 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-serving-cert\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.599118 master-0 kubenswrapper[24928]: I1205 10:47:36.599065 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.599118 master-0 kubenswrapper[24928]: I1205 10:47:36.599107 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.599214 master-0 kubenswrapper[24928]: I1205 10:47:36.599142 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24l8d\" (UniqueName: \"kubernetes.io/projected/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-kube-api-access-24l8d\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.599300 master-0 kubenswrapper[24928]: I1205 10:47:36.599244 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-dir\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.599346 master-0 kubenswrapper[24928]: I1205 10:47:36.599301 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-service-ca\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.599346 master-0 kubenswrapper[24928]: I1205 10:47:36.599333 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.599412 master-0 kubenswrapper[24928]: I1205 10:47:36.599359 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.599412 master-0 kubenswrapper[24928]: I1205 10:47:36.599392 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.599504 master-0 kubenswrapper[24928]: I1205 10:47:36.599448 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-policies\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.599504 master-0 kubenswrapper[24928]: I1205 10:47:36.599487 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-login\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.612279 master-0 kubenswrapper[24928]: I1205 10:47:36.612231 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-service-ca\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: I1205 10:47:36.612929 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-router-certs\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: I1205 10:47:36.612982 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-dir\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: I1205 10:47:36.613458 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: I1205 10:47:36.613700 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-login\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: I1205 10:47:36.613865 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-error\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: I1205 10:47:36.613864 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-serving-cert\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: E1205 10:47:36.614241 24928 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: configmap "v4-0-config-system-cliconfig" not found Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: E1205 10:47:36.614313 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig podName:75d79a3b-db06-4b9d-99ca-9b1a12dd44da nodeName:}" failed. No retries permitted until 2025-12-05 10:47:37.114293701 +0000 UTC m=+17.117487552 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig") pod "oauth-openshift-77b5b8969c-5clks" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da") : configmap "v4-0-config-system-cliconfig" not found Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: E1205 10:47:36.614665 24928 secret.go:189] Couldn't get secret openshift-authentication/v4-0-config-system-session: secret "v4-0-config-system-session" not found Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: E1205 10:47:36.614803 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session podName:75d79a3b-db06-4b9d-99ca-9b1a12dd44da nodeName:}" failed. No retries permitted until 2025-12-05 10:47:37.114776992 +0000 UTC m=+17.117970853 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session") pod "oauth-openshift-77b5b8969c-5clks" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da") : secret "v4-0-config-system-session" not found Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: I1205 10:47:36.615326 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: I1205 10:47:36.615550 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-policies\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.618084 master-0 kubenswrapper[24928]: I1205 10:47:36.617996 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:36.633839 master-0 kubenswrapper[24928]: I1205 10:47:36.633788 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24l8d\" (UniqueName: \"kubernetes.io/projected/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-kube-api-access-24l8d\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:37.208557 master-0 kubenswrapper[24928]: I1205 10:47:37.208505 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:37.208758 master-0 kubenswrapper[24928]: E1205 10:47:37.208631 24928 secret.go:189] Couldn't get secret openshift-authentication/v4-0-config-system-session: secret "v4-0-config-system-session" not found Dec 05 10:47:37.208796 master-0 kubenswrapper[24928]: E1205 10:47:37.208761 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session podName:75d79a3b-db06-4b9d-99ca-9b1a12dd44da nodeName:}" failed. No retries permitted until 2025-12-05 10:47:38.208745608 +0000 UTC m=+18.211939459 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session") pod "oauth-openshift-77b5b8969c-5clks" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da") : secret "v4-0-config-system-session" not found Dec 05 10:47:37.209006 master-0 kubenswrapper[24928]: E1205 10:47:37.208943 24928 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: configmap "v4-0-config-system-cliconfig" not found Dec 05 10:47:37.209086 master-0 kubenswrapper[24928]: I1205 10:47:37.209056 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:37.209200 master-0 kubenswrapper[24928]: E1205 10:47:37.209065 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig podName:75d79a3b-db06-4b9d-99ca-9b1a12dd44da nodeName:}" failed. No retries permitted until 2025-12-05 10:47:38.209041876 +0000 UTC m=+18.212235727 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig") pod "oauth-openshift-77b5b8969c-5clks" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da") : configmap "v4-0-config-system-cliconfig" not found Dec 05 10:47:37.696498 master-0 kubenswrapper[24928]: I1205 10:47:37.696416 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:47:37.752724 master-0 kubenswrapper[24928]: I1205 10:47:37.752654 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6p8cq" Dec 05 10:47:38.223111 master-0 kubenswrapper[24928]: I1205 10:47:38.223033 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:38.223344 master-0 kubenswrapper[24928]: I1205 10:47:38.223169 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:38.223743 master-0 kubenswrapper[24928]: E1205 10:47:38.223402 24928 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: configmap "v4-0-config-system-cliconfig" not found Dec 05 10:47:38.223743 master-0 kubenswrapper[24928]: E1205 10:47:38.223592 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig podName:75d79a3b-db06-4b9d-99ca-9b1a12dd44da nodeName:}" failed. No retries permitted until 2025-12-05 10:47:40.223548398 +0000 UTC m=+20.226742249 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig") pod "oauth-openshift-77b5b8969c-5clks" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da") : configmap "v4-0-config-system-cliconfig" not found Dec 05 10:47:38.224195 master-0 kubenswrapper[24928]: E1205 10:47:38.224165 24928 secret.go:189] Couldn't get secret openshift-authentication/v4-0-config-system-session: secret "v4-0-config-system-session" not found Dec 05 10:47:38.224265 master-0 kubenswrapper[24928]: E1205 10:47:38.224205 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session podName:75d79a3b-db06-4b9d-99ca-9b1a12dd44da nodeName:}" failed. No retries permitted until 2025-12-05 10:47:40.224194504 +0000 UTC m=+20.227388355 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session") pod "oauth-openshift-77b5b8969c-5clks" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da") : secret "v4-0-config-system-session" not found Dec 05 10:47:38.582278 master-0 kubenswrapper[24928]: I1205 10:47:38.582202 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-master-0_a906debd0c35952850935aee2d607cce/startup-monitor/0.log" Dec 05 10:47:38.582526 master-0 kubenswrapper[24928]: I1205 10:47:38.582284 24928 generic.go:334] "Generic (PLEG): container finished" podID="a906debd0c35952850935aee2d607cce" containerID="8930f7af011c1e43a64ba6a4ca85659538a4774816c777a7c34e7a4b10c3cab3" exitCode=137 Dec 05 10:47:38.934062 master-0 kubenswrapper[24928]: I1205 10:47:38.933998 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-master-0_a906debd0c35952850935aee2d607cce/startup-monitor/0.log" Dec 05 10:47:38.934666 master-0 kubenswrapper[24928]: I1205 10:47:38.934088 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:39.034680 master-0 kubenswrapper[24928]: I1205 10:47:39.034609 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-log\") pod \"a906debd0c35952850935aee2d607cce\" (UID: \"a906debd0c35952850935aee2d607cce\") " Dec 05 10:47:39.034921 master-0 kubenswrapper[24928]: I1205 10:47:39.034786 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-manifests\") pod \"a906debd0c35952850935aee2d607cce\" (UID: \"a906debd0c35952850935aee2d607cce\") " Dec 05 10:47:39.034921 master-0 kubenswrapper[24928]: I1205 10:47:39.034797 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-log" (OuterVolumeSpecName: "var-log") pod "a906debd0c35952850935aee2d607cce" (UID: "a906debd0c35952850935aee2d607cce"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:39.034921 master-0 kubenswrapper[24928]: I1205 10:47:39.034863 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-resource-dir\") pod \"a906debd0c35952850935aee2d607cce\" (UID: \"a906debd0c35952850935aee2d607cce\") " Dec 05 10:47:39.034921 master-0 kubenswrapper[24928]: I1205 10:47:39.034902 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-manifests" (OuterVolumeSpecName: "manifests") pod "a906debd0c35952850935aee2d607cce" (UID: "a906debd0c35952850935aee2d607cce"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:39.034921 master-0 kubenswrapper[24928]: I1205 10:47:39.034906 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-lock\") pod \"a906debd0c35952850935aee2d607cce\" (UID: \"a906debd0c35952850935aee2d607cce\") " Dec 05 10:47:39.035138 master-0 kubenswrapper[24928]: I1205 10:47:39.034966 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-lock" (OuterVolumeSpecName: "var-lock") pod "a906debd0c35952850935aee2d607cce" (UID: "a906debd0c35952850935aee2d607cce"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:39.035138 master-0 kubenswrapper[24928]: I1205 10:47:39.034985 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-pod-resource-dir\") pod \"a906debd0c35952850935aee2d607cce\" (UID: \"a906debd0c35952850935aee2d607cce\") " Dec 05 10:47:39.035138 master-0 kubenswrapper[24928]: I1205 10:47:39.035053 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "a906debd0c35952850935aee2d607cce" (UID: "a906debd0c35952850935aee2d607cce"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:39.035513 master-0 kubenswrapper[24928]: I1205 10:47:39.035477 24928 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-manifests\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:39.035513 master-0 kubenswrapper[24928]: I1205 10:47:39.035496 24928 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-resource-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:39.035513 master-0 kubenswrapper[24928]: I1205 10:47:39.035506 24928 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:39.035513 master-0 kubenswrapper[24928]: I1205 10:47:39.035517 24928 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-var-log\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:39.040672 master-0 kubenswrapper[24928]: I1205 10:47:39.040596 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "a906debd0c35952850935aee2d607cce" (UID: "a906debd0c35952850935aee2d607cce"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:47:39.136593 master-0 kubenswrapper[24928]: I1205 10:47:39.136471 24928 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/a906debd0c35952850935aee2d607cce-pod-resource-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:39.601937 master-0 kubenswrapper[24928]: I1205 10:47:39.601883 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-master-0_a906debd0c35952850935aee2d607cce/startup-monitor/0.log" Dec 05 10:47:39.602140 master-0 kubenswrapper[24928]: I1205 10:47:39.601956 24928 scope.go:117] "RemoveContainer" containerID="8930f7af011c1e43a64ba6a4ca85659538a4774816c777a7c34e7a4b10c3cab3" Dec 05 10:47:39.602140 master-0 kubenswrapper[24928]: I1205 10:47:39.602105 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:47:39.639308 master-0 kubenswrapper[24928]: I1205 10:47:39.639236 24928 kubelet.go:2706] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" mirrorPodUID="8624a3a9-8f59-4163-bdd7-271d28c9125c" Dec 05 10:47:40.218551 master-0 kubenswrapper[24928]: I1205 10:47:40.218475 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a906debd0c35952850935aee2d607cce" path="/var/lib/kubelet/pods/a906debd0c35952850935aee2d607cce/volumes" Dec 05 10:47:40.219262 master-0 kubenswrapper[24928]: I1205 10:47:40.218930 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" podUID="" Dec 05 10:47:40.228751 master-0 kubenswrapper[24928]: I1205 10:47:40.228695 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0"] Dec 05 10:47:40.228751 master-0 kubenswrapper[24928]: I1205 10:47:40.228739 24928 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" mirrorPodUID="8624a3a9-8f59-4163-bdd7-271d28c9125c" Dec 05 10:47:40.231926 master-0 kubenswrapper[24928]: I1205 10:47:40.231871 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0"] Dec 05 10:47:40.231926 master-0 kubenswrapper[24928]: I1205 10:47:40.231917 24928 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" mirrorPodUID="8624a3a9-8f59-4163-bdd7-271d28c9125c" Dec 05 10:47:40.250130 master-0 kubenswrapper[24928]: I1205 10:47:40.250030 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:40.250469 master-0 kubenswrapper[24928]: I1205 10:47:40.250384 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:40.250573 master-0 kubenswrapper[24928]: E1205 10:47:40.250543 24928 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: configmap "v4-0-config-system-cliconfig" not found Dec 05 10:47:40.250651 master-0 kubenswrapper[24928]: E1205 10:47:40.250626 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig podName:75d79a3b-db06-4b9d-99ca-9b1a12dd44da nodeName:}" failed. No retries permitted until 2025-12-05 10:47:44.250606294 +0000 UTC m=+24.253800225 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig") pod "oauth-openshift-77b5b8969c-5clks" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da") : configmap "v4-0-config-system-cliconfig" not found Dec 05 10:47:40.254183 master-0 kubenswrapper[24928]: I1205 10:47:40.254145 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:41.465528 master-0 kubenswrapper[24928]: I1205 10:47:41.465455 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:41.466013 master-0 kubenswrapper[24928]: I1205 10:47:41.465561 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:41.466013 master-0 kubenswrapper[24928]: E1205 10:47:41.465740 24928 projected.go:288] Couldn't get configMap openshift-kube-apiserver/kube-root-ca.crt: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:41.466013 master-0 kubenswrapper[24928]: E1205 10:47:41.465764 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-apiserver/installer-3-master-0: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:41.466013 master-0 kubenswrapper[24928]: E1205 10:47:41.465786 24928 projected.go:288] Couldn't get configMap openshift-etcd/kube-root-ca.crt: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:41.466013 master-0 kubenswrapper[24928]: E1205 10:47:41.465872 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-etcd/installer-2-master-0: object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:41.466013 master-0 kubenswrapper[24928]: E1205 10:47:41.465826 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access podName:75ba20a1-b322-4ce8-88d2-d8ee210cc27a nodeName:}" failed. No retries permitted until 2025-12-05 10:47:57.465806261 +0000 UTC m=+37.469000112 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access") pod "installer-3-master-0" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a") : object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:41.466207 master-0 kubenswrapper[24928]: E1205 10:47:41.466099 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access podName:68058d25-5888-495f-8571-453f0bf919d9 nodeName:}" failed. No retries permitted until 2025-12-05 10:47:57.466018266 +0000 UTC m=+37.469212157 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access") pod "installer-2-master-0" (UID: "68058d25-5888-495f-8571-453f0bf919d9") : object "openshift-etcd"/"kube-root-ca.crt" not registered Dec 05 10:47:44.304567 master-0 kubenswrapper[24928]: I1205 10:47:44.304482 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:44.305222 master-0 kubenswrapper[24928]: E1205 10:47:44.304762 24928 configmap.go:193] Couldn't get configMap openshift-authentication/v4-0-config-system-cliconfig: configmap "v4-0-config-system-cliconfig" not found Dec 05 10:47:44.305222 master-0 kubenswrapper[24928]: E1205 10:47:44.304826 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig podName:75d79a3b-db06-4b9d-99ca-9b1a12dd44da nodeName:}" failed. No retries permitted until 2025-12-05 10:47:52.304807488 +0000 UTC m=+32.308001339 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "v4-0-config-system-cliconfig" (UniqueName: "kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig") pod "oauth-openshift-77b5b8969c-5clks" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da") : configmap "v4-0-config-system-cliconfig" not found Dec 05 10:47:45.797187 master-0 kubenswrapper[24928]: I1205 10:47:45.797116 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:45.800767 master-0 kubenswrapper[24928]: I1205 10:47:45.800733 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:47:51.810470 master-0 kubenswrapper[24928]: I1205 10:47:51.803994 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-54dbc87ccb-m7p5f"] Dec 05 10:47:51.810470 master-0 kubenswrapper[24928]: I1205 10:47:51.805072 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:51.810470 master-0 kubenswrapper[24928]: I1205 10:47:51.807072 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 10:47:51.810470 master-0 kubenswrapper[24928]: I1205 10:47:51.807143 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 10:47:51.810470 master-0 kubenswrapper[24928]: I1205 10:47:51.807292 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 10:47:51.810470 master-0 kubenswrapper[24928]: I1205 10:47:51.808359 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 10:47:51.821063 master-0 kubenswrapper[24928]: I1205 10:47:51.820954 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-54dbc87ccb-m7p5f"] Dec 05 10:47:51.821683 master-0 kubenswrapper[24928]: I1205 10:47:51.821638 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 10:47:51.920337 master-0 kubenswrapper[24928]: I1205 10:47:51.920106 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-trusted-ca\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:51.920337 master-0 kubenswrapper[24928]: I1205 10:47:51.920190 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-serving-cert\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:51.920337 master-0 kubenswrapper[24928]: I1205 10:47:51.920231 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-config\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:51.920337 master-0 kubenswrapper[24928]: I1205 10:47:51.920303 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mznbp\" (UniqueName: \"kubernetes.io/projected/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-kube-api-access-mznbp\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:52.021949 master-0 kubenswrapper[24928]: I1205 10:47:52.021881 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mznbp\" (UniqueName: \"kubernetes.io/projected/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-kube-api-access-mznbp\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:52.021949 master-0 kubenswrapper[24928]: I1205 10:47:52.021956 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-trusted-ca\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:52.022241 master-0 kubenswrapper[24928]: I1205 10:47:52.022011 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-serving-cert\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:52.022241 master-0 kubenswrapper[24928]: I1205 10:47:52.022041 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-config\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:52.022965 master-0 kubenswrapper[24928]: I1205 10:47:52.022929 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-config\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:52.024309 master-0 kubenswrapper[24928]: I1205 10:47:52.024267 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-trusted-ca\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:52.027307 master-0 kubenswrapper[24928]: I1205 10:47:52.027269 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-serving-cert\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:52.040202 master-0 kubenswrapper[24928]: I1205 10:47:52.040151 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mznbp\" (UniqueName: \"kubernetes.io/projected/f3fa06fd-e24e-4bb1-8d20-be16c535d89f-kube-api-access-mznbp\") pod \"console-operator-54dbc87ccb-m7p5f\" (UID: \"f3fa06fd-e24e-4bb1-8d20-be16c535d89f\") " pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:52.159621 master-0 kubenswrapper[24928]: I1205 10:47:52.159484 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:52.326468 master-0 kubenswrapper[24928]: I1205 10:47:52.326380 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:52.327568 master-0 kubenswrapper[24928]: I1205 10:47:52.327522 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig\") pod \"oauth-openshift-77b5b8969c-5clks\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:52.349563 master-0 kubenswrapper[24928]: I1205 10:47:52.349279 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:52.349563 master-0 kubenswrapper[24928]: I1205 10:47:52.349516 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:47:52.379142 master-0 kubenswrapper[24928]: I1205 10:47:52.379093 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rsfjs" Dec 05 10:47:52.389976 master-0 kubenswrapper[24928]: I1205 10:47:52.389894 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:52.563157 master-0 kubenswrapper[24928]: I1205 10:47:52.563097 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-54dbc87ccb-m7p5f"] Dec 05 10:47:52.581155 master-0 kubenswrapper[24928]: W1205 10:47:52.581075 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3fa06fd_e24e_4bb1_8d20_be16c535d89f.slice/crio-a4d4f12386d81117307956a468e64f2cbcce89c7f97cf7f947fad9548667367e WatchSource:0}: Error finding container a4d4f12386d81117307956a468e64f2cbcce89c7f97cf7f947fad9548667367e: Status 404 returned error can't find the container with id a4d4f12386d81117307956a468e64f2cbcce89c7f97cf7f947fad9548667367e Dec 05 10:47:52.585223 master-0 kubenswrapper[24928]: I1205 10:47:52.583261 24928 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:47:52.684332 master-0 kubenswrapper[24928]: I1205 10:47:52.684201 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" event={"ID":"f3fa06fd-e24e-4bb1-8d20-be16c535d89f","Type":"ContainerStarted","Data":"a4d4f12386d81117307956a468e64f2cbcce89c7f97cf7f947fad9548667367e"} Dec 05 10:47:52.858542 master-0 kubenswrapper[24928]: I1205 10:47:52.858446 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-77b5b8969c-5clks"] Dec 05 10:47:52.868628 master-0 kubenswrapper[24928]: W1205 10:47:52.868552 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75d79a3b_db06_4b9d_99ca_9b1a12dd44da.slice/crio-c4bda59ea4c50819f401aa22ef3a9aed48de657128588d26b3684173ed617432 WatchSource:0}: Error finding container c4bda59ea4c50819f401aa22ef3a9aed48de657128588d26b3684173ed617432: Status 404 returned error can't find the container with id c4bda59ea4c50819f401aa22ef3a9aed48de657128588d26b3684173ed617432 Dec 05 10:47:53.017621 master-0 kubenswrapper[24928]: I1205 10:47:53.017291 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/installer-2-retry-1-master-0"] Dec 05 10:47:53.018337 master-0 kubenswrapper[24928]: I1205 10:47:53.018311 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:47:53.023242 master-0 kubenswrapper[24928]: I1205 10:47:53.020746 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd"/"kube-root-ca.crt" Dec 05 10:47:53.023242 master-0 kubenswrapper[24928]: I1205 10:47:53.020792 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd"/"installer-sa-dockercfg-76wfq" Dec 05 10:47:53.030681 master-0 kubenswrapper[24928]: I1205 10:47:53.030618 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd/installer-2-retry-1-master-0"] Dec 05 10:47:53.044029 master-0 kubenswrapper[24928]: I1205 10:47:53.043964 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee0f69ec-1862-4016-bf9d-7edf6554f473-kube-api-access\") pod \"installer-2-retry-1-master-0\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:47:53.044240 master-0 kubenswrapper[24928]: I1205 10:47:53.044100 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-var-lock\") pod \"installer-2-retry-1-master-0\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:47:53.044240 master-0 kubenswrapper[24928]: I1205 10:47:53.044160 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-kubelet-dir\") pod \"installer-2-retry-1-master-0\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:47:53.145345 master-0 kubenswrapper[24928]: I1205 10:47:53.145278 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-var-lock\") pod \"installer-2-retry-1-master-0\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:47:53.145778 master-0 kubenswrapper[24928]: I1205 10:47:53.145373 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-var-lock\") pod \"installer-2-retry-1-master-0\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:47:53.145778 master-0 kubenswrapper[24928]: I1205 10:47:53.145439 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-kubelet-dir\") pod \"installer-2-retry-1-master-0\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:47:53.145778 master-0 kubenswrapper[24928]: I1205 10:47:53.145517 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-kubelet-dir\") pod \"installer-2-retry-1-master-0\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:47:53.145778 master-0 kubenswrapper[24928]: I1205 10:47:53.145519 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee0f69ec-1862-4016-bf9d-7edf6554f473-kube-api-access\") pod \"installer-2-retry-1-master-0\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:47:53.163649 master-0 kubenswrapper[24928]: I1205 10:47:53.162410 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee0f69ec-1862-4016-bf9d-7edf6554f473-kube-api-access\") pod \"installer-2-retry-1-master-0\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:47:53.368931 master-0 kubenswrapper[24928]: I1205 10:47:53.368809 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:47:53.599027 master-0 kubenswrapper[24928]: I1205 10:47:53.598933 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd/installer-2-retry-1-master-0"] Dec 05 10:47:53.690790 master-0 kubenswrapper[24928]: I1205 10:47:53.690733 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-2-retry-1-master-0" event={"ID":"ee0f69ec-1862-4016-bf9d-7edf6554f473","Type":"ContainerStarted","Data":"212a02c8d8d6be333290bcb926f389ffc396525305548d0c39cda260ae72468f"} Dec 05 10:47:53.693367 master-0 kubenswrapper[24928]: I1205 10:47:53.693306 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" event={"ID":"75d79a3b-db06-4b9d-99ca-9b1a12dd44da","Type":"ContainerStarted","Data":"c4bda59ea4c50819f401aa22ef3a9aed48de657128588d26b3684173ed617432"} Dec 05 10:47:54.073368 master-0 kubenswrapper[24928]: I1205 10:47:54.072838 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-77b5b8969c-5clks"] Dec 05 10:47:54.701497 master-0 kubenswrapper[24928]: I1205 10:47:54.701439 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-2-retry-1-master-0" event={"ID":"ee0f69ec-1862-4016-bf9d-7edf6554f473","Type":"ContainerStarted","Data":"65a2979f50126cee418c5079b957bacb2ab47de1f45bc884591b9006fd791955"} Dec 05 10:47:54.721285 master-0 kubenswrapper[24928]: I1205 10:47:54.721199 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/installer-2-retry-1-master-0" podStartSLOduration=1.721163139 podStartE2EDuration="1.721163139s" podCreationTimestamp="2025-12-05 10:47:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:47:54.715699496 +0000 UTC m=+34.718893357" watchObservedRunningTime="2025-12-05 10:47:54.721163139 +0000 UTC m=+34.724356990" Dec 05 10:47:55.225468 master-0 kubenswrapper[24928]: I1205 10:47:55.225412 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8"] Dec 05 10:47:55.229566 master-0 kubenswrapper[24928]: I1205 10:47:55.229531 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8" Dec 05 10:47:55.234377 master-0 kubenswrapper[24928]: I1205 10:47:55.231808 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-xntfn" Dec 05 10:47:55.234377 master-0 kubenswrapper[24928]: I1205 10:47:55.232386 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Dec 05 10:47:55.243308 master-0 kubenswrapper[24928]: I1205 10:47:55.242354 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8"] Dec 05 10:47:55.387009 master-0 kubenswrapper[24928]: I1205 10:47:55.386948 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/e6acb017-11a4-426a-ae9e-1590f23e8090-monitoring-plugin-cert\") pod \"monitoring-plugin-54d7d75457-2k7b8\" (UID: \"e6acb017-11a4-426a-ae9e-1590f23e8090\") " pod="openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8" Dec 05 10:47:55.493807 master-0 kubenswrapper[24928]: I1205 10:47:55.493662 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/e6acb017-11a4-426a-ae9e-1590f23e8090-monitoring-plugin-cert\") pod \"monitoring-plugin-54d7d75457-2k7b8\" (UID: \"e6acb017-11a4-426a-ae9e-1590f23e8090\") " pod="openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8" Dec 05 10:47:55.499679 master-0 kubenswrapper[24928]: I1205 10:47:55.497850 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/e6acb017-11a4-426a-ae9e-1590f23e8090-monitoring-plugin-cert\") pod \"monitoring-plugin-54d7d75457-2k7b8\" (UID: \"e6acb017-11a4-426a-ae9e-1590f23e8090\") " pod="openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8" Dec 05 10:47:55.547333 master-0 kubenswrapper[24928]: I1205 10:47:55.547213 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8" Dec 05 10:47:56.521105 master-0 kubenswrapper[24928]: I1205 10:47:56.521049 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8"] Dec 05 10:47:56.528918 master-0 kubenswrapper[24928]: W1205 10:47:56.528867 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6acb017_11a4_426a_ae9e_1590f23e8090.slice/crio-d9c2ba815455bae202cd9569c40e13fcded8fb4ddb11436d92b154ed6f122198 WatchSource:0}: Error finding container d9c2ba815455bae202cd9569c40e13fcded8fb4ddb11436d92b154ed6f122198: Status 404 returned error can't find the container with id d9c2ba815455bae202cd9569c40e13fcded8fb4ddb11436d92b154ed6f122198 Dec 05 10:47:56.713149 master-0 kubenswrapper[24928]: I1205 10:47:56.713003 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8" event={"ID":"e6acb017-11a4-426a-ae9e-1590f23e8090","Type":"ContainerStarted","Data":"d9c2ba815455bae202cd9569c40e13fcded8fb4ddb11436d92b154ed6f122198"} Dec 05 10:47:56.715362 master-0 kubenswrapper[24928]: I1205 10:47:56.715322 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" event={"ID":"75d79a3b-db06-4b9d-99ca-9b1a12dd44da","Type":"ContainerStarted","Data":"a3d9aac442ecac29f714aebcf4a9f8ed868e053cf80b56659fc043a5d4b052e9"} Dec 05 10:47:56.715632 master-0 kubenswrapper[24928]: I1205 10:47:56.715532 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:56.717903 master-0 kubenswrapper[24928]: I1205 10:47:56.717790 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" event={"ID":"f3fa06fd-e24e-4bb1-8d20-be16c535d89f","Type":"ContainerStarted","Data":"79ac0355b4d1f2cf0298e3aa55437e8dff1a69bf178adb9e188d2826a8a0ef7e"} Dec 05 10:47:56.718871 master-0 kubenswrapper[24928]: I1205 10:47:56.718846 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:56.742947 master-0 kubenswrapper[24928]: I1205 10:47:56.742877 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" podStartSLOduration=17.476807203 podStartE2EDuration="20.742846525s" podCreationTimestamp="2025-12-05 10:47:36 +0000 UTC" firstStartedPulling="2025-12-05 10:47:52.870397104 +0000 UTC m=+32.873590955" lastFinishedPulling="2025-12-05 10:47:56.136436406 +0000 UTC m=+36.139630277" observedRunningTime="2025-12-05 10:47:56.742839395 +0000 UTC m=+36.746033256" watchObservedRunningTime="2025-12-05 10:47:56.742846525 +0000 UTC m=+36.746040376" Dec 05 10:47:56.769797 master-0 kubenswrapper[24928]: I1205 10:47:56.769707 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" podStartSLOduration=2.189615463 podStartE2EDuration="5.769686198s" podCreationTimestamp="2025-12-05 10:47:51 +0000 UTC" firstStartedPulling="2025-12-05 10:47:52.583218744 +0000 UTC m=+32.586412595" lastFinishedPulling="2025-12-05 10:47:56.163289469 +0000 UTC m=+36.166483330" observedRunningTime="2025-12-05 10:47:56.7635847 +0000 UTC m=+36.766778551" watchObservedRunningTime="2025-12-05 10:47:56.769686198 +0000 UTC m=+36.772880049" Dec 05 10:47:56.775332 master-0 kubenswrapper[24928]: I1205 10:47:56.772499 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-54dbc87ccb-m7p5f" Dec 05 10:47:56.880746 master-0 kubenswrapper[24928]: I1205 10:47:56.880282 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:47:57.028977 master-0 kubenswrapper[24928]: I1205 10:47:57.028572 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-69cd4c69bf-d9jtn"] Dec 05 10:47:57.032467 master-0 kubenswrapper[24928]: I1205 10:47:57.029526 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-69cd4c69bf-d9jtn" Dec 05 10:47:57.037220 master-0 kubenswrapper[24928]: I1205 10:47:57.033762 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 10:47:57.037220 master-0 kubenswrapper[24928]: I1205 10:47:57.033876 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 10:47:57.082558 master-0 kubenswrapper[24928]: I1205 10:47:57.082415 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6flz6\" (UniqueName: \"kubernetes.io/projected/3c708092-6503-4a61-9230-617f8a1e2d19-kube-api-access-6flz6\") pod \"downloads-69cd4c69bf-d9jtn\" (UID: \"3c708092-6503-4a61-9230-617f8a1e2d19\") " pod="openshift-console/downloads-69cd4c69bf-d9jtn" Dec 05 10:47:57.085076 master-0 kubenswrapper[24928]: I1205 10:47:57.085033 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-69cd4c69bf-d9jtn"] Dec 05 10:47:57.184241 master-0 kubenswrapper[24928]: I1205 10:47:57.184191 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6flz6\" (UniqueName: \"kubernetes.io/projected/3c708092-6503-4a61-9230-617f8a1e2d19-kube-api-access-6flz6\") pod \"downloads-69cd4c69bf-d9jtn\" (UID: \"3c708092-6503-4a61-9230-617f8a1e2d19\") " pod="openshift-console/downloads-69cd4c69bf-d9jtn" Dec 05 10:47:57.202749 master-0 kubenswrapper[24928]: I1205 10:47:57.202706 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6flz6\" (UniqueName: \"kubernetes.io/projected/3c708092-6503-4a61-9230-617f8a1e2d19-kube-api-access-6flz6\") pod \"downloads-69cd4c69bf-d9jtn\" (UID: \"3c708092-6503-4a61-9230-617f8a1e2d19\") " pod="openshift-console/downloads-69cd4c69bf-d9jtn" Dec 05 10:47:57.403142 master-0 kubenswrapper[24928]: I1205 10:47:57.403086 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-69cd4c69bf-d9jtn" Dec 05 10:47:57.488019 master-0 kubenswrapper[24928]: I1205 10:47:57.487951 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:57.488124 master-0 kubenswrapper[24928]: I1205 10:47:57.488083 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:47:57.488237 master-0 kubenswrapper[24928]: E1205 10:47:57.488207 24928 projected.go:288] Couldn't get configMap openshift-kube-apiserver/kube-root-ca.crt: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:57.488237 master-0 kubenswrapper[24928]: E1205 10:47:57.488234 24928 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-apiserver/installer-3-master-0: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:57.488305 master-0 kubenswrapper[24928]: E1205 10:47:57.488294 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access podName:75ba20a1-b322-4ce8-88d2-d8ee210cc27a nodeName:}" failed. No retries permitted until 2025-12-05 10:48:29.488277028 +0000 UTC m=+69.491470889 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access") pod "installer-3-master-0" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a") : object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 05 10:47:57.493495 master-0 kubenswrapper[24928]: I1205 10:47:57.491381 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"installer-2-master-0\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " pod="openshift-etcd/installer-2-master-0" Dec 05 10:47:57.698610 master-0 kubenswrapper[24928]: I1205 10:47:57.692862 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") pod \"68058d25-5888-495f-8571-453f0bf919d9\" (UID: \"68058d25-5888-495f-8571-453f0bf919d9\") " Dec 05 10:47:57.705401 master-0 kubenswrapper[24928]: I1205 10:47:57.705049 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "68058d25-5888-495f-8571-453f0bf919d9" (UID: "68058d25-5888-495f-8571-453f0bf919d9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:47:57.791862 master-0 kubenswrapper[24928]: I1205 10:47:57.791780 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-69cd4c69bf-d9jtn"] Dec 05 10:47:57.794464 master-0 kubenswrapper[24928]: I1205 10:47:57.794412 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/68058d25-5888-495f-8571-453f0bf919d9-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:47:58.028702 master-0 kubenswrapper[24928]: W1205 10:47:58.028570 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c708092_6503_4a61_9230_617f8a1e2d19.slice/crio-b85ac91b08da2d5a9d56252834a372fcd6dea939d6a6a950359b0e036206dcde WatchSource:0}: Error finding container b85ac91b08da2d5a9d56252834a372fcd6dea939d6a6a950359b0e036206dcde: Status 404 returned error can't find the container with id b85ac91b08da2d5a9d56252834a372fcd6dea939d6a6a950359b0e036206dcde Dec 05 10:47:58.732848 master-0 kubenswrapper[24928]: I1205 10:47:58.732799 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8" event={"ID":"e6acb017-11a4-426a-ae9e-1590f23e8090","Type":"ContainerStarted","Data":"f20cf09cfd7668a7837a188ccb45fb1321c701f694c254b5634a7eb676d9010e"} Dec 05 10:47:58.733388 master-0 kubenswrapper[24928]: I1205 10:47:58.733010 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8" Dec 05 10:47:58.733889 master-0 kubenswrapper[24928]: I1205 10:47:58.733820 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-69cd4c69bf-d9jtn" event={"ID":"3c708092-6503-4a61-9230-617f8a1e2d19","Type":"ContainerStarted","Data":"b85ac91b08da2d5a9d56252834a372fcd6dea939d6a6a950359b0e036206dcde"} Dec 05 10:47:58.739666 master-0 kubenswrapper[24928]: I1205 10:47:58.739626 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8" Dec 05 10:47:58.750294 master-0 kubenswrapper[24928]: I1205 10:47:58.749706 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-54d7d75457-2k7b8" podStartSLOduration=2.207038982 podStartE2EDuration="3.749688909s" podCreationTimestamp="2025-12-05 10:47:55 +0000 UTC" firstStartedPulling="2025-12-05 10:47:56.531543362 +0000 UTC m=+36.534737213" lastFinishedPulling="2025-12-05 10:47:58.074193289 +0000 UTC m=+38.077387140" observedRunningTime="2025-12-05 10:47:58.748734366 +0000 UTC m=+38.751928227" watchObservedRunningTime="2025-12-05 10:47:58.749688909 +0000 UTC m=+38.752882750" Dec 05 10:48:00.519511 master-0 kubenswrapper[24928]: I1205 10:48:00.518860 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-74f96dcf4d-9gskd"] Dec 05 10:48:00.520140 master-0 kubenswrapper[24928]: I1205 10:48:00.519832 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.522929 master-0 kubenswrapper[24928]: I1205 10:48:00.522889 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 10:48:00.523198 master-0 kubenswrapper[24928]: I1205 10:48:00.523165 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-2bqcq" Dec 05 10:48:00.523674 master-0 kubenswrapper[24928]: I1205 10:48:00.523265 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 10:48:00.523674 master-0 kubenswrapper[24928]: I1205 10:48:00.523349 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 10:48:00.523674 master-0 kubenswrapper[24928]: I1205 10:48:00.523508 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 10:48:00.523674 master-0 kubenswrapper[24928]: I1205 10:48:00.523541 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 10:48:00.538577 master-0 kubenswrapper[24928]: I1205 10:48:00.538259 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74f96dcf4d-9gskd"] Dec 05 10:48:00.655454 master-0 kubenswrapper[24928]: I1205 10:48:00.653956 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-service-ca\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.655454 master-0 kubenswrapper[24928]: I1205 10:48:00.654066 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-oauth-serving-cert\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.655454 master-0 kubenswrapper[24928]: I1205 10:48:00.654256 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-oauth-config\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.655454 master-0 kubenswrapper[24928]: I1205 10:48:00.654359 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsgbq\" (UniqueName: \"kubernetes.io/projected/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-kube-api-access-nsgbq\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.655454 master-0 kubenswrapper[24928]: I1205 10:48:00.654463 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-serving-cert\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.655454 master-0 kubenswrapper[24928]: I1205 10:48:00.654645 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-config\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.756544 master-0 kubenswrapper[24928]: I1205 10:48:00.756486 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-config\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.756838 master-0 kubenswrapper[24928]: I1205 10:48:00.756568 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-service-ca\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.756838 master-0 kubenswrapper[24928]: I1205 10:48:00.756780 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-oauth-serving-cert\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.757109 master-0 kubenswrapper[24928]: I1205 10:48:00.757075 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-oauth-config\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.757376 master-0 kubenswrapper[24928]: I1205 10:48:00.757341 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsgbq\" (UniqueName: \"kubernetes.io/projected/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-kube-api-access-nsgbq\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.758225 master-0 kubenswrapper[24928]: I1205 10:48:00.757971 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-oauth-serving-cert\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.758291 master-0 kubenswrapper[24928]: I1205 10:48:00.758244 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-config\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.758404 master-0 kubenswrapper[24928]: I1205 10:48:00.758259 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-serving-cert\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.758537 master-0 kubenswrapper[24928]: I1205 10:48:00.758369 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-service-ca\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.761295 master-0 kubenswrapper[24928]: I1205 10:48:00.761244 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-oauth-config\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.761727 master-0 kubenswrapper[24928]: I1205 10:48:00.761692 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-serving-cert\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.774132 master-0 kubenswrapper[24928]: I1205 10:48:00.774041 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsgbq\" (UniqueName: \"kubernetes.io/projected/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-kube-api-access-nsgbq\") pod \"console-74f96dcf4d-9gskd\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:00.860114 master-0 kubenswrapper[24928]: I1205 10:48:00.860046 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:01.281331 master-0 kubenswrapper[24928]: I1205 10:48:01.281287 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74f96dcf4d-9gskd"] Dec 05 10:48:01.285916 master-0 kubenswrapper[24928]: W1205 10:48:01.285859 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3cf0f61e_27cf_4f7d_b8da_c2f8f88f6527.slice/crio-940d3650e9f9322807a81f39b2daf6291f7767c05f0fcc28414d6332cac43bcd WatchSource:0}: Error finding container 940d3650e9f9322807a81f39b2daf6291f7767c05f0fcc28414d6332cac43bcd: Status 404 returned error can't find the container with id 940d3650e9f9322807a81f39b2daf6291f7767c05f0fcc28414d6332cac43bcd Dec 05 10:48:01.757609 master-0 kubenswrapper[24928]: I1205 10:48:01.757547 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74f96dcf4d-9gskd" event={"ID":"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527","Type":"ContainerStarted","Data":"940d3650e9f9322807a81f39b2daf6291f7767c05f0fcc28414d6332cac43bcd"} Dec 05 10:48:04.556450 master-0 kubenswrapper[24928]: I1205 10:48:04.556327 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-79cdddb8b4-mwjwx"] Dec 05 10:48:04.557916 master-0 kubenswrapper[24928]: I1205 10:48:04.557696 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.567306 master-0 kubenswrapper[24928]: I1205 10:48:04.567205 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 10:48:04.572525 master-0 kubenswrapper[24928]: I1205 10:48:04.572457 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-79cdddb8b4-mwjwx"] Dec 05 10:48:04.724615 master-0 kubenswrapper[24928]: I1205 10:48:04.724552 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vmlc\" (UniqueName: \"kubernetes.io/projected/ac38876d-122b-4146-babf-722930c2c4ae-kube-api-access-7vmlc\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.724615 master-0 kubenswrapper[24928]: I1205 10:48:04.724608 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-console-config\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.725170 master-0 kubenswrapper[24928]: I1205 10:48:04.724688 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-service-ca\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.725170 master-0 kubenswrapper[24928]: I1205 10:48:04.724842 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-trusted-ca-bundle\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.725170 master-0 kubenswrapper[24928]: I1205 10:48:04.724882 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-oauth-serving-cert\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.725170 master-0 kubenswrapper[24928]: I1205 10:48:04.724936 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-oauth-config\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.725170 master-0 kubenswrapper[24928]: I1205 10:48:04.724983 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-serving-cert\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.826531 master-0 kubenswrapper[24928]: I1205 10:48:04.826388 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-trusted-ca-bundle\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.826531 master-0 kubenswrapper[24928]: I1205 10:48:04.826475 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-oauth-serving-cert\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.826531 master-0 kubenswrapper[24928]: I1205 10:48:04.826521 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-oauth-config\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.826785 master-0 kubenswrapper[24928]: I1205 10:48:04.826564 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-serving-cert\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.826785 master-0 kubenswrapper[24928]: I1205 10:48:04.826594 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vmlc\" (UniqueName: \"kubernetes.io/projected/ac38876d-122b-4146-babf-722930c2c4ae-kube-api-access-7vmlc\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.826785 master-0 kubenswrapper[24928]: I1205 10:48:04.826620 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-console-config\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.826785 master-0 kubenswrapper[24928]: I1205 10:48:04.826681 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-service-ca\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.827724 master-0 kubenswrapper[24928]: I1205 10:48:04.827682 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-service-ca\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.827951 master-0 kubenswrapper[24928]: I1205 10:48:04.827913 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-trusted-ca-bundle\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.828352 master-0 kubenswrapper[24928]: I1205 10:48:04.828313 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-oauth-serving-cert\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.829083 master-0 kubenswrapper[24928]: I1205 10:48:04.829031 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-console-config\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.832927 master-0 kubenswrapper[24928]: I1205 10:48:04.832871 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-serving-cert\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.833095 master-0 kubenswrapper[24928]: I1205 10:48:04.833057 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-oauth-config\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.847379 master-0 kubenswrapper[24928]: I1205 10:48:04.847308 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vmlc\" (UniqueName: \"kubernetes.io/projected/ac38876d-122b-4146-babf-722930c2c4ae-kube-api-access-7vmlc\") pod \"console-79cdddb8b4-mwjwx\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:04.902145 master-0 kubenswrapper[24928]: I1205 10:48:04.902029 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:06.101048 master-0 kubenswrapper[24928]: I1205 10:48:06.100973 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-79cdddb8b4-mwjwx"] Dec 05 10:48:06.107874 master-0 kubenswrapper[24928]: W1205 10:48:06.107798 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac38876d_122b_4146_babf_722930c2c4ae.slice/crio-232db71c9974b1ab9b26647298ec9c771d701ecd59647a4c717df6ca62c916c3 WatchSource:0}: Error finding container 232db71c9974b1ab9b26647298ec9c771d701ecd59647a4c717df6ca62c916c3: Status 404 returned error can't find the container with id 232db71c9974b1ab9b26647298ec9c771d701ecd59647a4c717df6ca62c916c3 Dec 05 10:48:06.815083 master-0 kubenswrapper[24928]: I1205 10:48:06.815034 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79cdddb8b4-mwjwx" event={"ID":"ac38876d-122b-4146-babf-722930c2c4ae","Type":"ContainerStarted","Data":"4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02"} Dec 05 10:48:06.815350 master-0 kubenswrapper[24928]: I1205 10:48:06.815334 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79cdddb8b4-mwjwx" event={"ID":"ac38876d-122b-4146-babf-722930c2c4ae","Type":"ContainerStarted","Data":"232db71c9974b1ab9b26647298ec9c771d701ecd59647a4c717df6ca62c916c3"} Dec 05 10:48:06.817712 master-0 kubenswrapper[24928]: I1205 10:48:06.817693 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74f96dcf4d-9gskd" event={"ID":"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527","Type":"ContainerStarted","Data":"61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2"} Dec 05 10:48:06.836327 master-0 kubenswrapper[24928]: I1205 10:48:06.836240 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-79cdddb8b4-mwjwx" podStartSLOduration=2.836225106 podStartE2EDuration="2.836225106s" podCreationTimestamp="2025-12-05 10:48:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:48:06.834306858 +0000 UTC m=+46.837500709" watchObservedRunningTime="2025-12-05 10:48:06.836225106 +0000 UTC m=+46.839418947" Dec 05 10:48:06.852676 master-0 kubenswrapper[24928]: I1205 10:48:06.852595 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-74f96dcf4d-9gskd" podStartSLOduration=2.44797017 podStartE2EDuration="6.852579993s" podCreationTimestamp="2025-12-05 10:48:00 +0000 UTC" firstStartedPulling="2025-12-05 10:48:01.289074775 +0000 UTC m=+41.292268626" lastFinishedPulling="2025-12-05 10:48:05.693684598 +0000 UTC m=+45.696878449" observedRunningTime="2025-12-05 10:48:06.851828535 +0000 UTC m=+46.855022406" watchObservedRunningTime="2025-12-05 10:48:06.852579993 +0000 UTC m=+46.855773844" Dec 05 10:48:10.861461 master-0 kubenswrapper[24928]: I1205 10:48:10.860771 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:10.861461 master-0 kubenswrapper[24928]: I1205 10:48:10.860921 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:48:10.864463 master-0 kubenswrapper[24928]: I1205 10:48:10.863864 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:48:10.864463 master-0 kubenswrapper[24928]: I1205 10:48:10.864006 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:48:13.296541 master-0 kubenswrapper[24928]: I1205 10:48:13.296275 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-4-master-0"] Dec 05 10:48:13.298184 master-0 kubenswrapper[24928]: I1205 10:48:13.298147 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:48:13.300505 master-0 kubenswrapper[24928]: I1205 10:48:13.300446 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-grfdw" Dec 05 10:48:13.300618 master-0 kubenswrapper[24928]: I1205 10:48:13.300530 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 10:48:13.326288 master-0 kubenswrapper[24928]: I1205 10:48:13.326107 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-4-master-0"] Dec 05 10:48:13.379519 master-0 kubenswrapper[24928]: I1205 10:48:13.379363 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kube-api-access\") pod \"installer-4-master-0\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:48:13.379736 master-0 kubenswrapper[24928]: I1205 10:48:13.379561 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kubelet-dir\") pod \"installer-4-master-0\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:48:13.379736 master-0 kubenswrapper[24928]: I1205 10:48:13.379590 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-var-lock\") pod \"installer-4-master-0\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:48:13.481547 master-0 kubenswrapper[24928]: I1205 10:48:13.481479 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kubelet-dir\") pod \"installer-4-master-0\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:48:13.481547 master-0 kubenswrapper[24928]: I1205 10:48:13.481540 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-var-lock\") pod \"installer-4-master-0\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:48:13.481850 master-0 kubenswrapper[24928]: I1205 10:48:13.481579 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kube-api-access\") pod \"installer-4-master-0\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:48:13.481850 master-0 kubenswrapper[24928]: I1205 10:48:13.481604 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-var-lock\") pod \"installer-4-master-0\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:48:13.481850 master-0 kubenswrapper[24928]: I1205 10:48:13.481573 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kubelet-dir\") pod \"installer-4-master-0\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:48:13.500163 master-0 kubenswrapper[24928]: I1205 10:48:13.500089 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kube-api-access\") pod \"installer-4-master-0\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:48:13.629821 master-0 kubenswrapper[24928]: I1205 10:48:13.629718 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:48:14.080954 master-0 kubenswrapper[24928]: I1205 10:48:14.080891 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-4-master-0"] Dec 05 10:48:14.872774 master-0 kubenswrapper[24928]: I1205 10:48:14.872718 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-4-master-0" event={"ID":"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3","Type":"ContainerStarted","Data":"edac655814a82eb2075cd815c3a2e8122778082bf81de28d1d2bd7119b304f24"} Dec 05 10:48:14.872774 master-0 kubenswrapper[24928]: I1205 10:48:14.872769 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-4-master-0" event={"ID":"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3","Type":"ContainerStarted","Data":"40d4380f5668ad51b46124ceac16f446bb22917608c39463ebc5f63eef4b59bd"} Dec 05 10:48:14.895164 master-0 kubenswrapper[24928]: I1205 10:48:14.889933 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-4-master-0" podStartSLOduration=1.889901272 podStartE2EDuration="1.889901272s" podCreationTimestamp="2025-12-05 10:48:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:48:14.888811106 +0000 UTC m=+54.892004967" watchObservedRunningTime="2025-12-05 10:48:14.889901272 +0000 UTC m=+54.893095123" Dec 05 10:48:14.903020 master-0 kubenswrapper[24928]: I1205 10:48:14.902955 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:14.903020 master-0 kubenswrapper[24928]: I1205 10:48:14.903023 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:48:14.904254 master-0 kubenswrapper[24928]: I1205 10:48:14.904206 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:48:14.904323 master-0 kubenswrapper[24928]: I1205 10:48:14.904259 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:48:20.155396 master-0 kubenswrapper[24928]: I1205 10:48:20.155263 24928 scope.go:117] "RemoveContainer" containerID="15c71c8a48ba2bacfb2bd2a3e4d67fbc729bbcb6ae32fc023ae61c538c7d05e9" Dec 05 10:48:20.861157 master-0 kubenswrapper[24928]: I1205 10:48:20.861097 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:48:20.861376 master-0 kubenswrapper[24928]: I1205 10:48:20.861168 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:48:21.741443 master-0 kubenswrapper[24928]: I1205 10:48:21.741336 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" podUID="75d79a3b-db06-4b9d-99ca-9b1a12dd44da" containerName="oauth-openshift" containerID="cri-o://a3d9aac442ecac29f714aebcf4a9f8ed868e053cf80b56659fc043a5d4b052e9" gracePeriod=15 Dec 05 10:48:21.924384 master-0 kubenswrapper[24928]: I1205 10:48:21.924281 24928 generic.go:334] "Generic (PLEG): container finished" podID="75d79a3b-db06-4b9d-99ca-9b1a12dd44da" containerID="a3d9aac442ecac29f714aebcf4a9f8ed868e053cf80b56659fc043a5d4b052e9" exitCode=0 Dec 05 10:48:21.924384 master-0 kubenswrapper[24928]: I1205 10:48:21.924334 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" event={"ID":"75d79a3b-db06-4b9d-99ca-9b1a12dd44da","Type":"ContainerDied","Data":"a3d9aac442ecac29f714aebcf4a9f8ed868e053cf80b56659fc043a5d4b052e9"} Dec 05 10:48:22.391067 master-0 kubenswrapper[24928]: I1205 10:48:22.390985 24928 patch_prober.go:28] interesting pod/oauth-openshift-77b5b8969c-5clks container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.128.0.81:6443/healthz\": dial tcp 10.128.0.81:6443: connect: connection refused" start-of-body= Dec 05 10:48:22.391067 master-0 kubenswrapper[24928]: I1205 10:48:22.391044 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" podUID="75d79a3b-db06-4b9d-99ca-9b1a12dd44da" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.128.0.81:6443/healthz\": dial tcp 10.128.0.81:6443: connect: connection refused" Dec 05 10:48:24.904336 master-0 kubenswrapper[24928]: I1205 10:48:24.904167 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:48:24.904336 master-0 kubenswrapper[24928]: I1205 10:48:24.904258 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:48:24.980558 master-0 kubenswrapper[24928]: I1205 10:48:24.980496 24928 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-etcd/etcd-master-0"] Dec 05 10:48:24.981873 master-0 kubenswrapper[24928]: I1205 10:48:24.981016 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-etcd/etcd-master-0" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcdctl" containerID="cri-o://50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab" gracePeriod=30 Dec 05 10:48:24.981873 master-0 kubenswrapper[24928]: I1205 10:48:24.981041 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-etcd/etcd-master-0" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-metrics" containerID="cri-o://11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5" gracePeriod=30 Dec 05 10:48:24.981873 master-0 kubenswrapper[24928]: I1205 10:48:24.981183 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-etcd/etcd-master-0" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-readyz" containerID="cri-o://74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526" gracePeriod=30 Dec 05 10:48:24.981873 master-0 kubenswrapper[24928]: I1205 10:48:24.980986 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-etcd/etcd-master-0" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-rev" containerID="cri-o://d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819" gracePeriod=30 Dec 05 10:48:24.981873 master-0 kubenswrapper[24928]: I1205 10:48:24.981140 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-etcd/etcd-master-0" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd" containerID="cri-o://27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e" gracePeriod=30 Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.983673 24928 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-master-0"] Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: E1205 10:48:24.983972 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24e01603234fe8003f8aae8171b0065" containerName="setup" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.983984 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24e01603234fe8003f8aae8171b0065" containerName="setup" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: E1205 10:48:24.983998 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-readyz" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984005 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-readyz" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: E1205 10:48:24.984014 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcdctl" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984020 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcdctl" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: E1205 10:48:24.984032 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-ensure-env-vars" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984038 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-ensure-env-vars" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: E1205 10:48:24.984047 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-resources-copy" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984054 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-resources-copy" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: E1205 10:48:24.984068 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-metrics" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984074 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-metrics" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: E1205 10:48:24.984085 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984090 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: E1205 10:48:24.984101 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-rev" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984107 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-rev" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984454 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-readyz" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984473 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-metrics" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984486 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-ensure-env-vars" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984495 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcdctl" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984503 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-rev" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984516 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984525 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24e01603234fe8003f8aae8171b0065" containerName="setup" Dec 05 10:48:24.984726 master-0 kubenswrapper[24928]: I1205 10:48:24.984534 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24e01603234fe8003f8aae8171b0065" containerName="etcd-resources-copy" Dec 05 10:48:25.154962 master-0 kubenswrapper[24928]: I1205 10:48:25.154848 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-resource-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.154962 master-0 kubenswrapper[24928]: I1205 10:48:25.154917 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-static-pod-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.154962 master-0 kubenswrapper[24928]: I1205 10:48:25.154953 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-log-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.155215 master-0 kubenswrapper[24928]: I1205 10:48:25.155148 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-data-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.155295 master-0 kubenswrapper[24928]: I1205 10:48:25.155255 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-usr-local-bin\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.155557 master-0 kubenswrapper[24928]: I1205 10:48:25.155528 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-cert-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.256373 master-0 kubenswrapper[24928]: I1205 10:48:25.256324 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-cert-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.256373 master-0 kubenswrapper[24928]: I1205 10:48:25.256382 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-resource-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.256687 master-0 kubenswrapper[24928]: I1205 10:48:25.256403 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-static-pod-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.256687 master-0 kubenswrapper[24928]: I1205 10:48:25.256445 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-log-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.256687 master-0 kubenswrapper[24928]: I1205 10:48:25.256475 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-data-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.256687 master-0 kubenswrapper[24928]: I1205 10:48:25.256495 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-usr-local-bin\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.256687 master-0 kubenswrapper[24928]: I1205 10:48:25.256607 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-usr-local-bin\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.256687 master-0 kubenswrapper[24928]: I1205 10:48:25.256643 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-cert-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.256687 master-0 kubenswrapper[24928]: I1205 10:48:25.256666 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-resource-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.256687 master-0 kubenswrapper[24928]: I1205 10:48:25.256689 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-static-pod-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.257033 master-0 kubenswrapper[24928]: I1205 10:48:25.256709 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-log-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.257033 master-0 kubenswrapper[24928]: I1205 10:48:25.256729 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/58d12e893528ad53a994f10901a644ea-data-dir\") pod \"etcd-master-0\" (UID: \"58d12e893528ad53a994f10901a644ea\") " pod="openshift-etcd/etcd-master-0" Dec 05 10:48:25.955341 master-0 kubenswrapper[24928]: I1205 10:48:25.955252 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd-rev/0.log" Dec 05 10:48:25.956299 master-0 kubenswrapper[24928]: I1205 10:48:25.956217 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd-metrics/0.log" Dec 05 10:48:25.958520 master-0 kubenswrapper[24928]: I1205 10:48:25.958371 24928 generic.go:334] "Generic (PLEG): container finished" podID="c24e01603234fe8003f8aae8171b0065" containerID="d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819" exitCode=2 Dec 05 10:48:25.958520 master-0 kubenswrapper[24928]: I1205 10:48:25.958473 24928 generic.go:334] "Generic (PLEG): container finished" podID="c24e01603234fe8003f8aae8171b0065" containerID="74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526" exitCode=0 Dec 05 10:48:25.958520 master-0 kubenswrapper[24928]: I1205 10:48:25.958489 24928 generic.go:334] "Generic (PLEG): container finished" podID="c24e01603234fe8003f8aae8171b0065" containerID="11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5" exitCode=2 Dec 05 10:48:25.967690 master-0 kubenswrapper[24928]: I1205 10:48:25.964909 24928 scope.go:117] "RemoveContainer" containerID="0c2a956260655c961495c5a173ae1d7bbf307d5b8972db5af276ccbded5567cc" Dec 05 10:48:25.984646 master-0 kubenswrapper[24928]: I1205 10:48:25.984521 24928 scope.go:117] "RemoveContainer" containerID="b51c377684aafed12957958fdc2a947c34fdd370a6c72a4ef4641c4738cf31b1" Dec 05 10:48:26.016958 master-0 kubenswrapper[24928]: I1205 10:48:26.016835 24928 scope.go:117] "RemoveContainer" containerID="f19decd176069e2f8d9129c470cba4222beb8bbea5c63430032cfc884efc92be" Dec 05 10:48:26.386237 master-0 kubenswrapper[24928]: I1205 10:48:26.385927 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473259 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-login\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473323 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24l8d\" (UniqueName: \"kubernetes.io/projected/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-kube-api-access-24l8d\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473350 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-ocp-branding-template\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473368 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473585 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-provider-selection\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473647 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-error\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473680 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-router-certs\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473725 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-policies\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473743 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473779 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-serving-cert\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473827 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-trusted-ca-bundle\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473846 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-service-ca\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.474049 master-0 kubenswrapper[24928]: I1205 10:48:26.473918 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-dir\") pod \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\" (UID: \"75d79a3b-db06-4b9d-99ca-9b1a12dd44da\") " Dec 05 10:48:26.475330 master-0 kubenswrapper[24928]: I1205 10:48:26.474311 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:48:26.475330 master-0 kubenswrapper[24928]: I1205 10:48:26.474678 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:48:26.475330 master-0 kubenswrapper[24928]: I1205 10:48:26.475076 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:48:26.475602 master-0 kubenswrapper[24928]: I1205 10:48:26.475344 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:48:26.475786 master-0 kubenswrapper[24928]: I1205 10:48:26.475715 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:48:26.476932 master-0 kubenswrapper[24928]: I1205 10:48:26.476805 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:48:26.476932 master-0 kubenswrapper[24928]: I1205 10:48:26.476875 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:48:26.477541 master-0 kubenswrapper[24928]: I1205 10:48:26.477463 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:48:26.477541 master-0 kubenswrapper[24928]: I1205 10:48:26.477480 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:48:26.478214 master-0 kubenswrapper[24928]: I1205 10:48:26.478160 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:48:26.478668 master-0 kubenswrapper[24928]: I1205 10:48:26.478546 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:48:26.478668 master-0 kubenswrapper[24928]: I1205 10:48:26.478617 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-kube-api-access-24l8d" (OuterVolumeSpecName: "kube-api-access-24l8d") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "kube-api-access-24l8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:48:26.479265 master-0 kubenswrapper[24928]: I1205 10:48:26.479197 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "75d79a3b-db06-4b9d-99ca-9b1a12dd44da" (UID: "75d79a3b-db06-4b9d-99ca-9b1a12dd44da"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:48:26.575745 master-0 kubenswrapper[24928]: I1205 10:48:26.575611 24928 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-trusted-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.575745 master-0 kubenswrapper[24928]: I1205 10:48:26.575701 24928 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-service-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.575745 master-0 kubenswrapper[24928]: I1205 10:48:26.575739 24928 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.576256 master-0 kubenswrapper[24928]: I1205 10:48:26.575775 24928 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-login\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.576256 master-0 kubenswrapper[24928]: I1205 10:48:26.575802 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24l8d\" (UniqueName: \"kubernetes.io/projected/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-kube-api-access-24l8d\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.576256 master-0 kubenswrapper[24928]: I1205 10:48:26.575829 24928 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-ocp-branding-template\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.576256 master-0 kubenswrapper[24928]: I1205 10:48:26.575860 24928 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-session\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.576256 master-0 kubenswrapper[24928]: I1205 10:48:26.575888 24928 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-provider-selection\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.576256 master-0 kubenswrapper[24928]: I1205 10:48:26.575916 24928 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-user-template-error\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.576256 master-0 kubenswrapper[24928]: I1205 10:48:26.575946 24928 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-router-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.576256 master-0 kubenswrapper[24928]: I1205 10:48:26.576007 24928 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-audit-policies\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.576256 master-0 kubenswrapper[24928]: I1205 10:48:26.576037 24928 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-cliconfig\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.576256 master-0 kubenswrapper[24928]: I1205 10:48:26.576067 24928 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75d79a3b-db06-4b9d-99ca-9b1a12dd44da-v4-0-config-system-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:26.965915 master-0 kubenswrapper[24928]: I1205 10:48:26.965806 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-69cd4c69bf-d9jtn" event={"ID":"3c708092-6503-4a61-9230-617f8a1e2d19","Type":"ContainerStarted","Data":"5370bf9c2193180c87da7e448879e2649331b1940608ea04415864bad2cb356f"} Dec 05 10:48:26.966675 master-0 kubenswrapper[24928]: I1205 10:48:26.966094 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-69cd4c69bf-d9jtn" Dec 05 10:48:26.967805 master-0 kubenswrapper[24928]: I1205 10:48:26.967754 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" event={"ID":"75d79a3b-db06-4b9d-99ca-9b1a12dd44da","Type":"ContainerDied","Data":"c4bda59ea4c50819f401aa22ef3a9aed48de657128588d26b3684173ed617432"} Dec 05 10:48:26.967805 master-0 kubenswrapper[24928]: I1205 10:48:26.967801 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" Dec 05 10:48:26.967805 master-0 kubenswrapper[24928]: I1205 10:48:26.967809 24928 scope.go:117] "RemoveContainer" containerID="a3d9aac442ecac29f714aebcf4a9f8ed868e053cf80b56659fc043a5d4b052e9" Dec 05 10:48:26.968490 master-0 kubenswrapper[24928]: I1205 10:48:26.968224 24928 patch_prober.go:28] interesting pod/downloads-69cd4c69bf-d9jtn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.128.0.85:8080/\": dial tcp 10.128.0.85:8080: connect: connection refused" start-of-body= Dec 05 10:48:26.968490 master-0 kubenswrapper[24928]: I1205 10:48:26.968304 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-69cd4c69bf-d9jtn" podUID="3c708092-6503-4a61-9230-617f8a1e2d19" containerName="download-server" probeResult="failure" output="Get \"http://10.128.0.85:8080/\": dial tcp 10.128.0.85:8080: connect: connection refused" Dec 05 10:48:27.403761 master-0 kubenswrapper[24928]: I1205 10:48:27.403649 24928 patch_prober.go:28] interesting pod/downloads-69cd4c69bf-d9jtn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.128.0.85:8080/\": dial tcp 10.128.0.85:8080: connect: connection refused" start-of-body= Dec 05 10:48:27.404108 master-0 kubenswrapper[24928]: I1205 10:48:27.403771 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-69cd4c69bf-d9jtn" podUID="3c708092-6503-4a61-9230-617f8a1e2d19" containerName="download-server" probeResult="failure" output="Get \"http://10.128.0.85:8080/\": dial tcp 10.128.0.85:8080: connect: connection refused" Dec 05 10:48:27.404108 master-0 kubenswrapper[24928]: I1205 10:48:27.403831 24928 patch_prober.go:28] interesting pod/downloads-69cd4c69bf-d9jtn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.128.0.85:8080/\": dial tcp 10.128.0.85:8080: connect: connection refused" start-of-body= Dec 05 10:48:27.404108 master-0 kubenswrapper[24928]: I1205 10:48:27.403901 24928 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-69cd4c69bf-d9jtn" podUID="3c708092-6503-4a61-9230-617f8a1e2d19" containerName="download-server" probeResult="failure" output="Get \"http://10.128.0.85:8080/\": dial tcp 10.128.0.85:8080: connect: connection refused" Dec 05 10:48:27.977630 master-0 kubenswrapper[24928]: I1205 10:48:27.977560 24928 patch_prober.go:28] interesting pod/downloads-69cd4c69bf-d9jtn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.128.0.85:8080/\": dial tcp 10.128.0.85:8080: connect: connection refused" start-of-body= Dec 05 10:48:27.978139 master-0 kubenswrapper[24928]: I1205 10:48:27.977634 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-69cd4c69bf-d9jtn" podUID="3c708092-6503-4a61-9230-617f8a1e2d19" containerName="download-server" probeResult="failure" output="Get \"http://10.128.0.85:8080/\": dial tcp 10.128.0.85:8080: connect: connection refused" Dec 05 10:48:29.524852 master-0 kubenswrapper[24928]: I1205 10:48:29.524773 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:48:29.527524 master-0 kubenswrapper[24928]: I1205 10:48:29.527490 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"installer-3-master-0\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " pod="openshift-kube-apiserver/installer-3-master-0" Dec 05 10:48:29.626211 master-0 kubenswrapper[24928]: I1205 10:48:29.626130 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") pod \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\" (UID: \"75ba20a1-b322-4ce8-88d2-d8ee210cc27a\") " Dec 05 10:48:29.629053 master-0 kubenswrapper[24928]: I1205 10:48:29.628977 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "75ba20a1-b322-4ce8-88d2-d8ee210cc27a" (UID: "75ba20a1-b322-4ce8-88d2-d8ee210cc27a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:48:29.728744 master-0 kubenswrapper[24928]: I1205 10:48:29.728616 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/75ba20a1-b322-4ce8-88d2-d8ee210cc27a-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:30.861238 master-0 kubenswrapper[24928]: I1205 10:48:30.861176 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:48:30.861861 master-0 kubenswrapper[24928]: I1205 10:48:30.861251 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:48:34.903690 master-0 kubenswrapper[24928]: I1205 10:48:34.903603 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:48:34.903690 master-0 kubenswrapper[24928]: I1205 10:48:34.903683 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:48:36.699450 master-0 kubenswrapper[24928]: E1205 10:48:36.699325 24928 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:48:37.413267 master-0 kubenswrapper[24928]: I1205 10:48:37.413170 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-69cd4c69bf-d9jtn" Dec 05 10:48:37.787541 master-0 kubenswrapper[24928]: E1205 10:48:37.787220 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:48:27Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:48:27Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:48:27Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:48:27Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b\\\"],\\\"sizeBytes\\\":1631758507},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:15adb3b2133604b064893f8009a74145e4c8bb5b134d111346dcccbdd2aa9bc2\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:164fc35a19aa6cc886c8015c8ee3eba4895e76b1152cb9d795e4f3154a8533a3\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1610512706},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9014f384de5f9a0b7418d5869ad349abb9588d16bd09ed650a163c045315dbff\\\"],\\\"sizeBytes\\\":1232140918},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:610b8d322265b2c9d6b07efb2be26bf4d91e428b46412d73f5bdae0218004794\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:eafb9c83c480396c34e85d1f5f5c2623be6305031245be36455850c0398bfcc7\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1209064267},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0029526507396e493c5dce1652c41ed9c239b29e84ee579a2735fdb1aa3bce83\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:e1d263cd2113e0727021ccf27c8a671f8cfeaefbf93d60e3a918d6f60c136c30\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201604946},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:e8990432556acad31519b1a73ec32f32d27c2034cf9e5cc4db8980efc7331594\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:ebe9f523f5c211a3a0f2570331dddcd5be15b12c1fecd9b8b121f881bfaad029\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1129027903},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6\\\"],\\\"sizeBytes\\\":983705650},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\\\"],\\\"sizeBytes\\\":938303566},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:01d2e67fd74086da701c39dac5b821822351cb0151f9afe72821c05df19953ad\\\"],\\\"sizeBytes\\\":912722556},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:61664aa69b33349cc6de45e44ae6033e7f483c034ea01c0d9a8ca08a12d88e3a\\\"],\\\"sizeBytes\\\":874825223},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:631a3798b749fecc041a99929eb946618df723e15055e805ff752a1a1273481c\\\"],\\\"sizeBytes\\\":870567329},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\\\"],\\\"sizeBytes\\\":857069957},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c2431a990bcddde98829abda81950247021a2ebbabc964b1516ea046b5f1d4e\\\"],\\\"sizeBytes\\\":856659740},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b12f830c3316aa4dc061c2d00c74126282b3e2bcccc301eab00d57fff3c4c7c\\\"],\\\"sizeBytes\\\":767284906},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cb3ec61f9a932a9ad13bdeb44bcf9477a8d5f728151d7f19ed3ef7d4b02b3a82\\\"],\\\"sizeBytes\\\":682371258},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:916566bb9d0143352324233d460ad94697719c11c8c9158e3aea8f475941751f\\\"],\\\"sizeBytes\\\":677523572},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b\\\"],\\\"sizeBytes\\\":672407260},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e27a636083db9043e3e4bbdc336b5e7fb5693422246e443fd1d913e157f01d46\\\"],\\\"sizeBytes\\\":628330376},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9724d2036305cbd729e1f484c5bad89971de977fff8a6723fef1873858dd1123\\\"],\\\"sizeBytes\\\":616108962},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:df606f3b71d4376d1a2108c09f0d3dab455fc30bcb67c60e91590c105e9025bf\\\"],\\\"sizeBytes\\\":583836304},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2\\\"],\\\"sizeBytes\\\":576619763},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8\\\"],\\\"sizeBytes\\\":552673986},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737\\\"],\\\"sizeBytes\\\":551889548},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc\\\"],\\\"sizeBytes\\\":543227406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718\\\"],\\\"sizeBytes\\\":532719167},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cfde59e48cd5dee3721f34d249cb119cc3259fd857965d34f9c7ed83b0c363a1\\\"],\\\"sizeBytes\\\":532402162},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f4d4282cb53325e737ad68abbfcb70687ae04fb50353f4f0ba0ba5703b15009a\\\"],\\\"sizeBytes\\\":512838054},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\\\"],\\\"sizeBytes\\\":512452153},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660\\\"],\\\"sizeBytes\\\":509437356},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:97d26892192b552c16527bf2771e1b86528ab581a02dd9279cdf71c194830e3e\\\"],\\\"sizeBytes\\\":508042119},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df\\\"],\\\"sizeBytes\\\":507687221},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8375671da86aa527ee7e291d86971b0baa823ffc7663b5a983084456e76c0f59\\\"],\\\"sizeBytes\\\":506741476},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0c3d16a01c2d60f9b536ca815ed8dc6abdca2b78e392551dc3fb79be537a354\\\"],\\\"sizeBytes\\\":506703191},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b\\\"],\\\"sizeBytes\\\":505649178},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:86af77350cfe6fd69280157e4162aa0147873d9431c641ae4ad3e881ff768a73\\\"],\\\"sizeBytes\\\":505628211},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\\\"],\\\"sizeBytes\\\":503340749},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8139ed65c0a0a4b0f253b715c11cc52be027efe8a4774da9ccce35c78ef439da\\\"],\\\"sizeBytes\\\":503011144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4\\\"],\\\"sizeBytes\\\":502436444},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10e57ca7611f79710f05777dc6a8f31c7e04eb09da4d8d793a5acfbf0e4692d7\\\"],\\\"sizeBytes\\\":500943492},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce\\\"],\\\"sizeBytes\\\":500848684},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c\\\"],\\\"sizeBytes\\\":499798563},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96\\\"],\\\"sizeBytes\\\":499705918},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33a20002692769235e95271ab071783c57ff50681088fa1035b86af31e73cf20\\\"],\\\"sizeBytes\\\":499125567},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:75d996f6147edb88c09fd1a052099de66638590d7d03a735006244bc9e19f898\\\"],\\\"sizeBytes\\\":499082775},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3b8d91a25eeb9f02041e947adb3487da3e7ab8449d3d2ad015827e7954df7b34\\\"],\\\"sizeBytes\\\":490455952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68\\\"],\\\"sizeBytes\\\":489528665},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1\\\"],\\\"sizeBytes\\\":481559117},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2b3d313c599852b3543ee5c3a62691bd2d1bbad12c2e1c610cd71a1dec6eea32\\\"],\\\"sizeBytes\\\":481499222},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3a77aa4d03b89ea284e3467a268e5989a77a2ef63e685eb1d5c5ea5b3922b7a\\\"],\\\"sizeBytes\\\":478917802},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eb928c13a46d3fb45f4a881892d023a92d610a5430be0ffd916aaf8da8e7d297\\\"],\\\"sizeBytes\\\":478642572}]}}\" for node \"master-0\": Patch \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0/status?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:48:40.861357 master-0 kubenswrapper[24928]: I1205 10:48:40.861215 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:48:40.861357 master-0 kubenswrapper[24928]: I1205 10:48:40.861306 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:48:41.408608 master-0 kubenswrapper[24928]: I1205 10:48:41.408466 24928 patch_prober.go:28] interesting pod/kube-controller-manager-master-0 container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.32.10:10257/healthz\": dial tcp 192.168.32.10:10257: connect: connection refused" start-of-body= Dec 05 10:48:41.408608 master-0 kubenswrapper[24928]: I1205 10:48:41.408542 24928 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.32.10:10257/healthz\": dial tcp 192.168.32.10:10257: connect: connection refused" Dec 05 10:48:42.080381 master-0 kubenswrapper[24928]: I1205 10:48:42.080309 24928 generic.go:334] "Generic (PLEG): container finished" podID="ee0f69ec-1862-4016-bf9d-7edf6554f473" containerID="65a2979f50126cee418c5079b957bacb2ab47de1f45bc884591b9006fd791955" exitCode=0 Dec 05 10:48:42.081016 master-0 kubenswrapper[24928]: I1205 10:48:42.080670 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-2-retry-1-master-0" event={"ID":"ee0f69ec-1862-4016-bf9d-7edf6554f473","Type":"ContainerDied","Data":"65a2979f50126cee418c5079b957bacb2ab47de1f45bc884591b9006fd791955"} Dec 05 10:48:42.083278 master-0 kubenswrapper[24928]: I1205 10:48:42.083239 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/0.log" Dec 05 10:48:42.083349 master-0 kubenswrapper[24928]: I1205 10:48:42.083286 24928 generic.go:334] "Generic (PLEG): container finished" podID="5219435a07a0220d41da97c4fb70abb1" containerID="a0c4bf77b56c9bf53793c3092e77173ade30a59f4de69720ced91f80c7a365c6" exitCode=1 Dec 05 10:48:42.083349 master-0 kubenswrapper[24928]: I1205 10:48:42.083309 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerDied","Data":"a0c4bf77b56c9bf53793c3092e77173ade30a59f4de69720ced91f80c7a365c6"} Dec 05 10:48:42.083848 master-0 kubenswrapper[24928]: I1205 10:48:42.083812 24928 scope.go:117] "RemoveContainer" containerID="a0c4bf77b56c9bf53793c3092e77173ade30a59f4de69720ced91f80c7a365c6" Dec 05 10:48:43.298376 master-0 kubenswrapper[24928]: I1205 10:48:43.298237 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:48:43.464402 master-0 kubenswrapper[24928]: I1205 10:48:43.464326 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:48:43.521214 master-0 kubenswrapper[24928]: I1205 10:48:43.521119 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-kubelet-dir\") pod \"ee0f69ec-1862-4016-bf9d-7edf6554f473\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " Dec 05 10:48:43.521491 master-0 kubenswrapper[24928]: I1205 10:48:43.521246 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-var-lock\") pod \"ee0f69ec-1862-4016-bf9d-7edf6554f473\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " Dec 05 10:48:43.521491 master-0 kubenswrapper[24928]: I1205 10:48:43.521246 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ee0f69ec-1862-4016-bf9d-7edf6554f473" (UID: "ee0f69ec-1862-4016-bf9d-7edf6554f473"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:48:43.521491 master-0 kubenswrapper[24928]: I1205 10:48:43.521347 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-var-lock" (OuterVolumeSpecName: "var-lock") pod "ee0f69ec-1862-4016-bf9d-7edf6554f473" (UID: "ee0f69ec-1862-4016-bf9d-7edf6554f473"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:48:43.521491 master-0 kubenswrapper[24928]: I1205 10:48:43.521400 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee0f69ec-1862-4016-bf9d-7edf6554f473-kube-api-access\") pod \"ee0f69ec-1862-4016-bf9d-7edf6554f473\" (UID: \"ee0f69ec-1862-4016-bf9d-7edf6554f473\") " Dec 05 10:48:43.521821 master-0 kubenswrapper[24928]: I1205 10:48:43.521735 24928 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:43.521821 master-0 kubenswrapper[24928]: I1205 10:48:43.521768 24928 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/ee0f69ec-1862-4016-bf9d-7edf6554f473-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:43.526824 master-0 kubenswrapper[24928]: I1205 10:48:43.526758 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee0f69ec-1862-4016-bf9d-7edf6554f473-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ee0f69ec-1862-4016-bf9d-7edf6554f473" (UID: "ee0f69ec-1862-4016-bf9d-7edf6554f473"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:48:43.622758 master-0 kubenswrapper[24928]: I1205 10:48:43.622698 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ee0f69ec-1862-4016-bf9d-7edf6554f473-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:44.108782 master-0 kubenswrapper[24928]: I1205 10:48:44.108700 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/installer-2-retry-1-master-0" event={"ID":"ee0f69ec-1862-4016-bf9d-7edf6554f473","Type":"ContainerDied","Data":"212a02c8d8d6be333290bcb926f389ffc396525305548d0c39cda260ae72468f"} Dec 05 10:48:44.109146 master-0 kubenswrapper[24928]: I1205 10:48:44.108790 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/installer-2-retry-1-master-0" Dec 05 10:48:44.109257 master-0 kubenswrapper[24928]: I1205 10:48:44.109236 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="212a02c8d8d6be333290bcb926f389ffc396525305548d0c39cda260ae72468f" Dec 05 10:48:44.904595 master-0 kubenswrapper[24928]: I1205 10:48:44.904529 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:48:44.905559 master-0 kubenswrapper[24928]: I1205 10:48:44.905502 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:48:45.121515 master-0 kubenswrapper[24928]: I1205 10:48:45.121414 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/0.log" Dec 05 10:48:45.121515 master-0 kubenswrapper[24928]: I1205 10:48:45.121507 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"cf784a713d53bc0541dfbf03c915a4f6b70148e2c7b42b695d4ce6650a26f08f"} Dec 05 10:48:46.700529 master-0 kubenswrapper[24928]: E1205 10:48:46.700387 24928 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:48:47.788273 master-0 kubenswrapper[24928]: E1205 10:48:47.788176 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:48:50.861318 master-0 kubenswrapper[24928]: I1205 10:48:50.861228 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:48:50.861318 master-0 kubenswrapper[24928]: I1205 10:48:50.861294 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:48:51.303764 master-0 kubenswrapper[24928]: I1205 10:48:51.303690 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:48:51.307616 master-0 kubenswrapper[24928]: I1205 10:48:51.307568 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:48:52.167331 master-0 kubenswrapper[24928]: I1205 10:48:52.167249 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:48:54.905470 master-0 kubenswrapper[24928]: I1205 10:48:54.905386 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:48:54.905978 master-0 kubenswrapper[24928]: I1205 10:48:54.905486 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:48:56.701494 master-0 kubenswrapper[24928]: E1205 10:48:56.701327 24928 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:48:57.788547 master-0 kubenswrapper[24928]: E1205 10:48:57.788414 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": context deadline exceeded" Dec 05 10:48:59.050980 master-0 kubenswrapper[24928]: E1205 10:48:59.050790 24928 event.go:359] "Server rejected event (will not retry!)" err="Timeout: request did not complete within requested timeout - context deadline exceeded" event="&Event{ObjectMeta:{etcd-master-0.187e4c08e037cc73 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-master-0,UID:c24e01603234fe8003f8aae8171b0065,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Killing,Message:Stopping container etcd-metrics,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:48:24.981015667 +0000 UTC m=+64.984209518,LastTimestamp:2025-12-05 10:48:24.981015667 +0000 UTC m=+64.984209518,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:48:59.062580 master-0 kubenswrapper[24928]: I1205 10:48:59.062467 24928 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.781065008s: [/var/lib/containers/storage/overlay/838a2b4c90cc2f830abe86eeef5093fe6609c1cc122b344ae6d897a555f8a0ec/diff /var/log/pods/openshift-console-operator_console-operator-54dbc87ccb-m7p5f_f3fa06fd-e24e-4bb1-8d20-be16c535d89f/console-operator/0.log]; will not log again for this container unless duration exceeds 2s Dec 05 10:48:59.220192 master-0 kubenswrapper[24928]: I1205 10:48:59.220135 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd-rev/0.log" Dec 05 10:48:59.221146 master-0 kubenswrapper[24928]: I1205 10:48:59.221104 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd-metrics/0.log" Dec 05 10:48:59.644535 master-0 kubenswrapper[24928]: I1205 10:48:59.644498 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd-rev/0.log" Dec 05 10:48:59.645213 master-0 kubenswrapper[24928]: I1205 10:48:59.645182 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd-metrics/0.log" Dec 05 10:48:59.645683 master-0 kubenswrapper[24928]: I1205 10:48:59.645650 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd/0.log" Dec 05 10:48:59.646084 master-0 kubenswrapper[24928]: I1205 10:48:59.646052 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcdctl/0.log" Dec 05 10:48:59.647303 master-0 kubenswrapper[24928]: I1205 10:48:59.647278 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-master-0" Dec 05 10:48:59.734653 master-0 kubenswrapper[24928]: I1205 10:48:59.734502 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcd/0.log" Dec 05 10:48:59.735861 master-0 kubenswrapper[24928]: I1205 10:48:59.735830 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_c24e01603234fe8003f8aae8171b0065/etcdctl/0.log" Dec 05 10:48:59.736983 master-0 kubenswrapper[24928]: I1205 10:48:59.736954 24928 generic.go:334] "Generic (PLEG): container finished" podID="c24e01603234fe8003f8aae8171b0065" containerID="27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e" exitCode=137 Dec 05 10:48:59.736983 master-0 kubenswrapper[24928]: I1205 10:48:59.736980 24928 generic.go:334] "Generic (PLEG): container finished" podID="c24e01603234fe8003f8aae8171b0065" containerID="50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab" exitCode=137 Dec 05 10:48:59.737104 master-0 kubenswrapper[24928]: I1205 10:48:59.737019 24928 scope.go:117] "RemoveContainer" containerID="d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819" Dec 05 10:48:59.751068 master-0 kubenswrapper[24928]: I1205 10:48:59.751022 24928 scope.go:117] "RemoveContainer" containerID="74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526" Dec 05 10:48:59.767372 master-0 kubenswrapper[24928]: I1205 10:48:59.767309 24928 scope.go:117] "RemoveContainer" containerID="11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5" Dec 05 10:48:59.783978 master-0 kubenswrapper[24928]: I1205 10:48:59.783905 24928 scope.go:117] "RemoveContainer" containerID="27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e" Dec 05 10:48:59.784120 master-0 kubenswrapper[24928]: I1205 10:48:59.784014 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-data-dir" (OuterVolumeSpecName: "data-dir") pod "c24e01603234fe8003f8aae8171b0065" (UID: "c24e01603234fe8003f8aae8171b0065"). InnerVolumeSpecName "data-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:48:59.784120 master-0 kubenswrapper[24928]: I1205 10:48:59.783961 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-data-dir\") pod \"c24e01603234fe8003f8aae8171b0065\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " Dec 05 10:48:59.784291 master-0 kubenswrapper[24928]: I1205 10:48:59.784260 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-log-dir\") pod \"c24e01603234fe8003f8aae8171b0065\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " Dec 05 10:48:59.784368 master-0 kubenswrapper[24928]: I1205 10:48:59.784299 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-usr-local-bin\") pod \"c24e01603234fe8003f8aae8171b0065\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " Dec 05 10:48:59.784368 master-0 kubenswrapper[24928]: I1205 10:48:59.784348 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-cert-dir\") pod \"c24e01603234fe8003f8aae8171b0065\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " Dec 05 10:48:59.784503 master-0 kubenswrapper[24928]: I1205 10:48:59.784390 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-static-pod-dir\") pod \"c24e01603234fe8003f8aae8171b0065\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " Dec 05 10:48:59.784503 master-0 kubenswrapper[24928]: I1205 10:48:59.784397 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-log-dir" (OuterVolumeSpecName: "log-dir") pod "c24e01603234fe8003f8aae8171b0065" (UID: "c24e01603234fe8003f8aae8171b0065"). InnerVolumeSpecName "log-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:48:59.784503 master-0 kubenswrapper[24928]: I1205 10:48:59.784438 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-resource-dir\") pod \"c24e01603234fe8003f8aae8171b0065\" (UID: \"c24e01603234fe8003f8aae8171b0065\") " Dec 05 10:48:59.784503 master-0 kubenswrapper[24928]: I1205 10:48:59.784465 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "c24e01603234fe8003f8aae8171b0065" (UID: "c24e01603234fe8003f8aae8171b0065"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:48:59.784503 master-0 kubenswrapper[24928]: I1205 10:48:59.784491 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-usr-local-bin" (OuterVolumeSpecName: "usr-local-bin") pod "c24e01603234fe8003f8aae8171b0065" (UID: "c24e01603234fe8003f8aae8171b0065"). InnerVolumeSpecName "usr-local-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:48:59.784503 master-0 kubenswrapper[24928]: I1205 10:48:59.784513 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-static-pod-dir" (OuterVolumeSpecName: "static-pod-dir") pod "c24e01603234fe8003f8aae8171b0065" (UID: "c24e01603234fe8003f8aae8171b0065"). InnerVolumeSpecName "static-pod-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:48:59.784748 master-0 kubenswrapper[24928]: I1205 10:48:59.784605 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "c24e01603234fe8003f8aae8171b0065" (UID: "c24e01603234fe8003f8aae8171b0065"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:48:59.785202 master-0 kubenswrapper[24928]: I1205 10:48:59.785162 24928 reconciler_common.go:293] "Volume detached for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-static-pod-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:59.785202 master-0 kubenswrapper[24928]: I1205 10:48:59.785189 24928 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-resource-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:59.785202 master-0 kubenswrapper[24928]: I1205 10:48:59.785200 24928 reconciler_common.go:293] "Volume detached for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-data-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:59.785345 master-0 kubenswrapper[24928]: I1205 10:48:59.785210 24928 reconciler_common.go:293] "Volume detached for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-log-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:59.785345 master-0 kubenswrapper[24928]: I1205 10:48:59.785221 24928 reconciler_common.go:293] "Volume detached for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-usr-local-bin\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:59.785345 master-0 kubenswrapper[24928]: I1205 10:48:59.785230 24928 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/c24e01603234fe8003f8aae8171b0065-cert-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:48:59.800856 master-0 kubenswrapper[24928]: I1205 10:48:59.800797 24928 scope.go:117] "RemoveContainer" containerID="50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab" Dec 05 10:48:59.817705 master-0 kubenswrapper[24928]: I1205 10:48:59.817642 24928 scope.go:117] "RemoveContainer" containerID="41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2" Dec 05 10:48:59.841483 master-0 kubenswrapper[24928]: I1205 10:48:59.840711 24928 scope.go:117] "RemoveContainer" containerID="d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc" Dec 05 10:48:59.861354 master-0 kubenswrapper[24928]: I1205 10:48:59.861295 24928 scope.go:117] "RemoveContainer" containerID="f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2" Dec 05 10:48:59.883340 master-0 kubenswrapper[24928]: I1205 10:48:59.883297 24928 scope.go:117] "RemoveContainer" containerID="d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819" Dec 05 10:48:59.883883 master-0 kubenswrapper[24928]: E1205 10:48:59.883837 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819\": container with ID starting with d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819 not found: ID does not exist" containerID="d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819" Dec 05 10:48:59.883956 master-0 kubenswrapper[24928]: I1205 10:48:59.883886 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819"} err="failed to get container status \"d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819\": rpc error: code = NotFound desc = could not find container \"d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819\": container with ID starting with d7266e051745d535e36951f78681c838e30c0a53c9a8b82f667ce9387b8e8819 not found: ID does not exist" Dec 05 10:48:59.883956 master-0 kubenswrapper[24928]: I1205 10:48:59.883940 24928 scope.go:117] "RemoveContainer" containerID="74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526" Dec 05 10:48:59.884354 master-0 kubenswrapper[24928]: E1205 10:48:59.884296 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526\": container with ID starting with 74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526 not found: ID does not exist" containerID="74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526" Dec 05 10:48:59.884456 master-0 kubenswrapper[24928]: I1205 10:48:59.884350 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526"} err="failed to get container status \"74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526\": rpc error: code = NotFound desc = could not find container \"74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526\": container with ID starting with 74d989b1f09ae6f454043a0ee9a45315d00031b51ad1626141cb15fb3f6cc526 not found: ID does not exist" Dec 05 10:48:59.884456 master-0 kubenswrapper[24928]: I1205 10:48:59.884378 24928 scope.go:117] "RemoveContainer" containerID="11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5" Dec 05 10:48:59.884783 master-0 kubenswrapper[24928]: E1205 10:48:59.884699 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5\": container with ID starting with 11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5 not found: ID does not exist" containerID="11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5" Dec 05 10:48:59.884783 master-0 kubenswrapper[24928]: I1205 10:48:59.884734 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5"} err="failed to get container status \"11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5\": rpc error: code = NotFound desc = could not find container \"11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5\": container with ID starting with 11533e5c2a0f19d5c1b0d9aee625d04005e795f5eddb8457ae179c3aa8d6f5f5 not found: ID does not exist" Dec 05 10:48:59.884919 master-0 kubenswrapper[24928]: I1205 10:48:59.884787 24928 scope.go:117] "RemoveContainer" containerID="27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e" Dec 05 10:48:59.885287 master-0 kubenswrapper[24928]: E1205 10:48:59.885250 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e\": container with ID starting with 27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e not found: ID does not exist" containerID="27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e" Dec 05 10:48:59.885287 master-0 kubenswrapper[24928]: I1205 10:48:59.885278 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e"} err="failed to get container status \"27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e\": rpc error: code = NotFound desc = could not find container \"27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e\": container with ID starting with 27a52ef6720d4bb8bfd6e41df7c7d401f3c8eb3aba0a71e3701b80830bbe2c0e not found: ID does not exist" Dec 05 10:48:59.885397 master-0 kubenswrapper[24928]: I1205 10:48:59.885298 24928 scope.go:117] "RemoveContainer" containerID="50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab" Dec 05 10:48:59.885716 master-0 kubenswrapper[24928]: E1205 10:48:59.885669 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab\": container with ID starting with 50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab not found: ID does not exist" containerID="50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab" Dec 05 10:48:59.885789 master-0 kubenswrapper[24928]: I1205 10:48:59.885715 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab"} err="failed to get container status \"50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab\": rpc error: code = NotFound desc = could not find container \"50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab\": container with ID starting with 50671b00dc1c685ace52b506ff62825564d9435dc1b26e5468ce69df89e66dab not found: ID does not exist" Dec 05 10:48:59.885789 master-0 kubenswrapper[24928]: I1205 10:48:59.885744 24928 scope.go:117] "RemoveContainer" containerID="41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2" Dec 05 10:48:59.886075 master-0 kubenswrapper[24928]: E1205 10:48:59.886020 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2\": container with ID starting with 41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2 not found: ID does not exist" containerID="41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2" Dec 05 10:48:59.886075 master-0 kubenswrapper[24928]: I1205 10:48:59.886052 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2"} err="failed to get container status \"41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2\": rpc error: code = NotFound desc = could not find container \"41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2\": container with ID starting with 41cb0dd82ecb0642a7c012cd90efa006658697eb3cfb61bd53943f8e4bbba9d2 not found: ID does not exist" Dec 05 10:48:59.886075 master-0 kubenswrapper[24928]: I1205 10:48:59.886069 24928 scope.go:117] "RemoveContainer" containerID="d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc" Dec 05 10:48:59.886578 master-0 kubenswrapper[24928]: E1205 10:48:59.886540 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc\": container with ID starting with d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc not found: ID does not exist" containerID="d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc" Dec 05 10:48:59.886661 master-0 kubenswrapper[24928]: I1205 10:48:59.886575 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc"} err="failed to get container status \"d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc\": rpc error: code = NotFound desc = could not find container \"d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc\": container with ID starting with d3beec3d29e6a74e30f873e6193c9761d4935d9c2905bbed7422ab63198ea6dc not found: ID does not exist" Dec 05 10:48:59.886661 master-0 kubenswrapper[24928]: I1205 10:48:59.886620 24928 scope.go:117] "RemoveContainer" containerID="f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2" Dec 05 10:48:59.887034 master-0 kubenswrapper[24928]: E1205 10:48:59.886996 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2\": container with ID starting with f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2 not found: ID does not exist" containerID="f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2" Dec 05 10:48:59.887034 master-0 kubenswrapper[24928]: I1205 10:48:59.887023 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2"} err="failed to get container status \"f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2\": rpc error: code = NotFound desc = could not find container \"f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2\": container with ID starting with f13bd1640c2dc1a7ee482e5db79cceea05252edf7e6e516b9105fcf44c6657d2 not found: ID does not exist" Dec 05 10:49:00.212723 master-0 kubenswrapper[24928]: I1205 10:49:00.212651 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c24e01603234fe8003f8aae8171b0065" path="/var/lib/kubelet/pods/c24e01603234fe8003f8aae8171b0065/volumes" Dec 05 10:49:00.745217 master-0 kubenswrapper[24928]: I1205 10:49:00.745149 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-4-master-0_f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3/installer/0.log" Dec 05 10:49:00.745519 master-0 kubenswrapper[24928]: I1205 10:49:00.745223 24928 generic.go:334] "Generic (PLEG): container finished" podID="f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3" containerID="edac655814a82eb2075cd815c3a2e8122778082bf81de28d1d2bd7119b304f24" exitCode=1 Dec 05 10:49:00.745519 master-0 kubenswrapper[24928]: I1205 10:49:00.745304 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-4-master-0" event={"ID":"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3","Type":"ContainerDied","Data":"edac655814a82eb2075cd815c3a2e8122778082bf81de28d1d2bd7119b304f24"} Dec 05 10:49:00.747726 master-0 kubenswrapper[24928]: I1205 10:49:00.747661 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-master-0" Dec 05 10:49:00.861216 master-0 kubenswrapper[24928]: I1205 10:49:00.861148 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:49:00.861216 master-0 kubenswrapper[24928]: I1205 10:49:00.861212 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:49:02.067074 master-0 kubenswrapper[24928]: I1205 10:49:02.067018 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-4-master-0_f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3/installer/0.log" Dec 05 10:49:02.067574 master-0 kubenswrapper[24928]: I1205 10:49:02.067101 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:49:02.114396 master-0 kubenswrapper[24928]: I1205 10:49:02.113939 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-var-lock\") pod \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " Dec 05 10:49:02.114396 master-0 kubenswrapper[24928]: I1205 10:49:02.114072 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kubelet-dir\") pod \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " Dec 05 10:49:02.114396 master-0 kubenswrapper[24928]: I1205 10:49:02.114106 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-var-lock" (OuterVolumeSpecName: "var-lock") pod "f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3" (UID: "f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:49:02.114396 master-0 kubenswrapper[24928]: I1205 10:49:02.114124 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kube-api-access\") pod \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\" (UID: \"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3\") " Dec 05 10:49:02.114396 master-0 kubenswrapper[24928]: I1205 10:49:02.114217 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3" (UID: "f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:49:02.115077 master-0 kubenswrapper[24928]: I1205 10:49:02.114727 24928 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:49:02.115077 master-0 kubenswrapper[24928]: I1205 10:49:02.114753 24928 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:49:02.117746 master-0 kubenswrapper[24928]: I1205 10:49:02.117651 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3" (UID: "f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:49:02.203637 master-0 kubenswrapper[24928]: I1205 10:49:02.203540 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-master-0" Dec 05 10:49:02.216950 master-0 kubenswrapper[24928]: I1205 10:49:02.216862 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:49:02.226314 master-0 kubenswrapper[24928]: I1205 10:49:02.226229 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:49:02.226314 master-0 kubenswrapper[24928]: I1205 10:49:02.226288 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:49:02.761672 master-0 kubenswrapper[24928]: I1205 10:49:02.761606 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-4-master-0_f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3/installer/0.log" Dec 05 10:49:02.761979 master-0 kubenswrapper[24928]: I1205 10:49:02.761684 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-4-master-0" event={"ID":"f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3","Type":"ContainerDied","Data":"40d4380f5668ad51b46124ceac16f446bb22917608c39463ebc5f63eef4b59bd"} Dec 05 10:49:02.761979 master-0 kubenswrapper[24928]: I1205 10:49:02.761714 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40d4380f5668ad51b46124ceac16f446bb22917608c39463ebc5f63eef4b59bd" Dec 05 10:49:02.761979 master-0 kubenswrapper[24928]: I1205 10:49:02.761782 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-4-master-0" Dec 05 10:49:03.302457 master-0 kubenswrapper[24928]: I1205 10:49:03.302356 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:49:04.903785 master-0 kubenswrapper[24928]: I1205 10:49:04.903697 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:49:04.904893 master-0 kubenswrapper[24928]: I1205 10:49:04.903778 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:49:06.702683 master-0 kubenswrapper[24928]: E1205 10:49:06.702585 24928 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:49:07.789164 master-0 kubenswrapper[24928]: E1205 10:49:07.789037 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:49:10.862630 master-0 kubenswrapper[24928]: I1205 10:49:10.862552 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:49:10.863359 master-0 kubenswrapper[24928]: I1205 10:49:10.862660 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:49:11.835100 master-0 kubenswrapper[24928]: I1205 10:49:11.835041 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-node-identity_network-node-identity-ql7j7_d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/approver/1.log" Dec 05 10:49:11.836638 master-0 kubenswrapper[24928]: I1205 10:49:11.836592 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-node-identity_network-node-identity-ql7j7_d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/approver/0.log" Dec 05 10:49:11.837175 master-0 kubenswrapper[24928]: I1205 10:49:11.837116 24928 generic.go:334] "Generic (PLEG): container finished" podID="d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e" containerID="75e16cb5a0af404fec070efd073424497756db4fbc0649e23d5c8b08a868d67f" exitCode=1 Dec 05 10:49:11.837291 master-0 kubenswrapper[24928]: I1205 10:49:11.837197 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-ql7j7" event={"ID":"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e","Type":"ContainerDied","Data":"75e16cb5a0af404fec070efd073424497756db4fbc0649e23d5c8b08a868d67f"} Dec 05 10:49:11.837291 master-0 kubenswrapper[24928]: I1205 10:49:11.837278 24928 scope.go:117] "RemoveContainer" containerID="5674d03e599731519f29f243f076ae28159025706182191e04e4bbfef0819512" Dec 05 10:49:11.838366 master-0 kubenswrapper[24928]: I1205 10:49:11.838319 24928 scope.go:117] "RemoveContainer" containerID="75e16cb5a0af404fec070efd073424497756db4fbc0649e23d5c8b08a868d67f" Dec 05 10:49:12.844867 master-0 kubenswrapper[24928]: I1205 10:49:12.844817 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-node-identity_network-node-identity-ql7j7_d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e/approver/1.log" Dec 05 10:49:12.845723 master-0 kubenswrapper[24928]: I1205 10:49:12.845680 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-ql7j7" event={"ID":"d807d0da-5cac-4c6a-afd8-5a1cd1e5e59e","Type":"ContainerStarted","Data":"4808a75b8478e4cacd32d1b42fa4036f20414a0cda55a2667be42248d038a571"} Dec 05 10:49:14.903948 master-0 kubenswrapper[24928]: I1205 10:49:14.903843 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:49:14.903948 master-0 kubenswrapper[24928]: I1205 10:49:14.903929 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:49:16.703539 master-0 kubenswrapper[24928]: E1205 10:49:16.703389 24928 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:49:16.703539 master-0 kubenswrapper[24928]: I1205 10:49:16.703519 24928 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 05 10:49:17.789561 master-0 kubenswrapper[24928]: E1205 10:49:17.789407 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:49:17.790409 master-0 kubenswrapper[24928]: E1205 10:49:17.790303 24928 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 10:49:20.861871 master-0 kubenswrapper[24928]: I1205 10:49:20.861809 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:49:20.862475 master-0 kubenswrapper[24928]: I1205 10:49:20.861899 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:49:24.904571 master-0 kubenswrapper[24928]: I1205 10:49:24.904487 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:49:24.905307 master-0 kubenswrapper[24928]: I1205 10:49:24.904576 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:49:26.387378 master-0 kubenswrapper[24928]: I1205 10:49:26.387272 24928 status_manager.go:851] "Failed to get status for pod" podUID="75d79a3b-db06-4b9d-99ca-9b1a12dd44da" pod="openshift-authentication/oauth-openshift-77b5b8969c-5clks" err="the server was unable to return a response in the time allotted, but may still be processing the request (get pods oauth-openshift-77b5b8969c-5clks)" Dec 05 10:49:26.704546 master-0 kubenswrapper[24928]: E1205 10:49:26.704380 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="200ms" Dec 05 10:49:30.861600 master-0 kubenswrapper[24928]: I1205 10:49:30.861538 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:49:30.862357 master-0 kubenswrapper[24928]: I1205 10:49:30.861612 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:49:33.065329 master-0 kubenswrapper[24928]: E1205 10:49:33.065104 24928 event.go:359] "Server rejected event (will not retry!)" err="Timeout: request did not complete within requested timeout - context deadline exceeded" event="&Event{ObjectMeta:{etcd-master-0.187e4c08e03933b8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-master-0,UID:c24e01603234fe8003f8aae8171b0065,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Killing,Message:Stopping container etcd,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:48:24.98110764 +0000 UTC m=+64.984301521,LastTimestamp:2025-12-05 10:48:24.98110764 +0000 UTC m=+64.984301521,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:49:34.904393 master-0 kubenswrapper[24928]: I1205 10:49:34.904311 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:49:34.904393 master-0 kubenswrapper[24928]: I1205 10:49:34.904381 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:49:36.230812 master-0 kubenswrapper[24928]: E1205 10:49:36.230749 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Timeout: request did not complete within requested timeout - context deadline exceeded" pod="openshift-etcd/etcd-master-0" Dec 05 10:49:36.232389 master-0 kubenswrapper[24928]: I1205 10:49:36.231162 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-master-0" Dec 05 10:49:36.905561 master-0 kubenswrapper[24928]: E1205 10:49:36.905500 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="400ms" Dec 05 10:49:37.038477 master-0 kubenswrapper[24928]: I1205 10:49:37.038341 24928 generic.go:334] "Generic (PLEG): container finished" podID="58d12e893528ad53a994f10901a644ea" containerID="43d3f4e122ed97183dd21d655df13c829e9b14d20b03aa1aeaca2312eec97b50" exitCode=0 Dec 05 10:49:37.038477 master-0 kubenswrapper[24928]: I1205 10:49:37.038467 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"58d12e893528ad53a994f10901a644ea","Type":"ContainerDied","Data":"43d3f4e122ed97183dd21d655df13c829e9b14d20b03aa1aeaca2312eec97b50"} Dec 05 10:49:37.039485 master-0 kubenswrapper[24928]: I1205 10:49:37.038522 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"58d12e893528ad53a994f10901a644ea","Type":"ContainerStarted","Data":"1edf3b6384b68eab6975bbc95d663ba47e1eec90da5aba61688ad6c819462fd7"} Dec 05 10:49:37.039485 master-0 kubenswrapper[24928]: I1205 10:49:37.039096 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:49:37.039485 master-0 kubenswrapper[24928]: I1205 10:49:37.039136 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:49:37.839967 master-0 kubenswrapper[24928]: E1205 10:49:37.839807 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:49:27Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:49:27Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:49:27Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:49:27Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:50e368e01772dd0dc9c4f9a6cdd5a9693a224968f75dc19eafe2a416f583bdab\\\"],\\\"sizeBytes\\\":2890347099},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b\\\"],\\\"sizeBytes\\\":1631758507},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:15adb3b2133604b064893f8009a74145e4c8bb5b134d111346dcccbdd2aa9bc2\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:164fc35a19aa6cc886c8015c8ee3eba4895e76b1152cb9d795e4f3154a8533a3\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1610512706},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9014f384de5f9a0b7418d5869ad349abb9588d16bd09ed650a163c045315dbff\\\"],\\\"sizeBytes\\\":1232140918},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:610b8d322265b2c9d6b07efb2be26bf4d91e428b46412d73f5bdae0218004794\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:eafb9c83c480396c34e85d1f5f5c2623be6305031245be36455850c0398bfcc7\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1209064267},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0029526507396e493c5dce1652c41ed9c239b29e84ee579a2735fdb1aa3bce83\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:e1d263cd2113e0727021ccf27c8a671f8cfeaefbf93d60e3a918d6f60c136c30\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201604946},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:e8990432556acad31519b1a73ec32f32d27c2034cf9e5cc4db8980efc7331594\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:ebe9f523f5c211a3a0f2570331dddcd5be15b12c1fecd9b8b121f881bfaad029\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1129027903},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6\\\"],\\\"sizeBytes\\\":983705650},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\\\"],\\\"sizeBytes\\\":938303566},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:01d2e67fd74086da701c39dac5b821822351cb0151f9afe72821c05df19953ad\\\"],\\\"sizeBytes\\\":912722556},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:61664aa69b33349cc6de45e44ae6033e7f483c034ea01c0d9a8ca08a12d88e3a\\\"],\\\"sizeBytes\\\":874825223},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:631a3798b749fecc041a99929eb946618df723e15055e805ff752a1a1273481c\\\"],\\\"sizeBytes\\\":870567329},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\\\"],\\\"sizeBytes\\\":857069957},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c2431a990bcddde98829abda81950247021a2ebbabc964b1516ea046b5f1d4e\\\"],\\\"sizeBytes\\\":856659740},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b12f830c3316aa4dc061c2d00c74126282b3e2bcccc301eab00d57fff3c4c7c\\\"],\\\"sizeBytes\\\":767284906},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cb3ec61f9a932a9ad13bdeb44bcf9477a8d5f728151d7f19ed3ef7d4b02b3a82\\\"],\\\"sizeBytes\\\":682371258},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:916566bb9d0143352324233d460ad94697719c11c8c9158e3aea8f475941751f\\\"],\\\"sizeBytes\\\":677523572},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b\\\"],\\\"sizeBytes\\\":672407260},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e27a636083db9043e3e4bbdc336b5e7fb5693422246e443fd1d913e157f01d46\\\"],\\\"sizeBytes\\\":628330376},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9724d2036305cbd729e1f484c5bad89971de977fff8a6723fef1873858dd1123\\\"],\\\"sizeBytes\\\":616108962},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:df606f3b71d4376d1a2108c09f0d3dab455fc30bcb67c60e91590c105e9025bf\\\"],\\\"sizeBytes\\\":583836304},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2\\\"],\\\"sizeBytes\\\":576619763},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8\\\"],\\\"sizeBytes\\\":552673986},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737\\\"],\\\"sizeBytes\\\":551889548},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc\\\"],\\\"sizeBytes\\\":543227406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718\\\"],\\\"sizeBytes\\\":532719167},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cfde59e48cd5dee3721f34d249cb119cc3259fd857965d34f9c7ed83b0c363a1\\\"],\\\"sizeBytes\\\":532402162},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f4d4282cb53325e737ad68abbfcb70687ae04fb50353f4f0ba0ba5703b15009a\\\"],\\\"sizeBytes\\\":512838054},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\\\"],\\\"sizeBytes\\\":512452153},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660\\\"],\\\"sizeBytes\\\":509437356},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:97d26892192b552c16527bf2771e1b86528ab581a02dd9279cdf71c194830e3e\\\"],\\\"sizeBytes\\\":508042119},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df\\\"],\\\"sizeBytes\\\":507687221},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8375671da86aa527ee7e291d86971b0baa823ffc7663b5a983084456e76c0f59\\\"],\\\"sizeBytes\\\":506741476},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0c3d16a01c2d60f9b536ca815ed8dc6abdca2b78e392551dc3fb79be537a354\\\"],\\\"sizeBytes\\\":506703191},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b\\\"],\\\"sizeBytes\\\":505649178},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:86af77350cfe6fd69280157e4162aa0147873d9431c641ae4ad3e881ff768a73\\\"],\\\"sizeBytes\\\":505628211},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\\\"],\\\"sizeBytes\\\":503340749},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8139ed65c0a0a4b0f253b715c11cc52be027efe8a4774da9ccce35c78ef439da\\\"],\\\"sizeBytes\\\":503011144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4\\\"],\\\"sizeBytes\\\":502436444},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10e57ca7611f79710f05777dc6a8f31c7e04eb09da4d8d793a5acfbf0e4692d7\\\"],\\\"sizeBytes\\\":500943492},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce\\\"],\\\"sizeBytes\\\":500848684},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c\\\"],\\\"sizeBytes\\\":499798563},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96\\\"],\\\"sizeBytes\\\":499705918},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33a20002692769235e95271ab071783c57ff50681088fa1035b86af31e73cf20\\\"],\\\"sizeBytes\\\":499125567},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:75d996f6147edb88c09fd1a052099de66638590d7d03a735006244bc9e19f898\\\"],\\\"sizeBytes\\\":499082775},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3b8d91a25eeb9f02041e947adb3487da3e7ab8449d3d2ad015827e7954df7b34\\\"],\\\"sizeBytes\\\":490455952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68\\\"],\\\"sizeBytes\\\":489528665},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1\\\"],\\\"sizeBytes\\\":481559117},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2b3d313c599852b3543ee5c3a62691bd2d1bbad12c2e1c610cd71a1dec6eea32\\\"],\\\"sizeBytes\\\":481499222},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3a77aa4d03b89ea284e3467a268e5989a77a2ef63e685eb1d5c5ea5b3922b7a\\\"],\\\"sizeBytes\\\":478917802}]}}\" for node \"master-0\": Patch \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0/status?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:49:40.861011 master-0 kubenswrapper[24928]: I1205 10:49:40.860899 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:49:40.861011 master-0 kubenswrapper[24928]: I1205 10:49:40.860976 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:49:44.904038 master-0 kubenswrapper[24928]: I1205 10:49:44.903939 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:49:44.905088 master-0 kubenswrapper[24928]: I1205 10:49:44.904683 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:49:47.306701 master-0 kubenswrapper[24928]: E1205 10:49:47.306593 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" interval="800ms" Dec 05 10:49:47.840841 master-0 kubenswrapper[24928]: E1205 10:49:47.840661 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:49:50.861019 master-0 kubenswrapper[24928]: I1205 10:49:50.860917 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:49:50.861019 master-0 kubenswrapper[24928]: I1205 10:49:50.860990 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:49:54.904169 master-0 kubenswrapper[24928]: I1205 10:49:54.904093 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:49:54.904885 master-0 kubenswrapper[24928]: I1205 10:49:54.904176 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:49:57.841052 master-0 kubenswrapper[24928]: E1205 10:49:57.840952 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:49:58.109262 master-0 kubenswrapper[24928]: E1205 10:49:58.109097 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="1.6s" Dec 05 10:50:00.861144 master-0 kubenswrapper[24928]: I1205 10:50:00.861070 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:50:00.861144 master-0 kubenswrapper[24928]: I1205 10:50:00.861131 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:50:04.903665 master-0 kubenswrapper[24928]: I1205 10:50:04.903607 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:50:04.903665 master-0 kubenswrapper[24928]: I1205 10:50:04.903668 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:50:07.067620 master-0 kubenswrapper[24928]: E1205 10:50:07.067399 24928 event.go:359] "Server rejected event (will not retry!)" err="Timeout: request did not complete within requested timeout - context deadline exceeded" event="&Event{ObjectMeta:{etcd-master-0.187e4c08e03a4065 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-master-0,UID:c24e01603234fe8003f8aae8171b0065,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Killing,Message:Stopping container etcd-readyz,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:48:24.981176421 +0000 UTC m=+64.984370272,LastTimestamp:2025-12-05 10:48:24.981176421 +0000 UTC m=+64.984370272,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:50:07.842093 master-0 kubenswrapper[24928]: E1205 10:50:07.841963 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:50:09.709717 master-0 kubenswrapper[24928]: E1205 10:50:09.709560 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 05 10:50:10.861153 master-0 kubenswrapper[24928]: I1205 10:50:10.861088 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:50:10.861809 master-0 kubenswrapper[24928]: I1205 10:50:10.861176 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:50:11.042733 master-0 kubenswrapper[24928]: E1205 10:50:11.042647 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Timeout: request did not complete within requested timeout - context deadline exceeded" pod="openshift-etcd/etcd-master-0" Dec 05 10:50:12.307145 master-0 kubenswrapper[24928]: I1205 10:50:12.307063 24928 generic.go:334] "Generic (PLEG): container finished" podID="926263c4-ec5b-41cb-9c30-0c88f636035f" containerID="9426824799d7ff29210670da649e79b8fec18b4165425c4e34958dec644e0eb2" exitCode=0 Dec 05 10:50:12.308040 master-0 kubenswrapper[24928]: I1205 10:50:12.307161 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" event={"ID":"926263c4-ec5b-41cb-9c30-0c88f636035f","Type":"ContainerDied","Data":"9426824799d7ff29210670da649e79b8fec18b4165425c4e34958dec644e0eb2"} Dec 05 10:50:12.308040 master-0 kubenswrapper[24928]: I1205 10:50:12.307224 24928 scope.go:117] "RemoveContainer" containerID="3504f06691ba2b115136a2091ae95118383b7c2cc711e4550b1e60138be21510" Dec 05 10:50:12.308040 master-0 kubenswrapper[24928]: I1205 10:50:12.307994 24928 scope.go:117] "RemoveContainer" containerID="9426824799d7ff29210670da649e79b8fec18b4165425c4e34958dec644e0eb2" Dec 05 10:50:12.311653 master-0 kubenswrapper[24928]: I1205 10:50:12.311557 24928 generic.go:334] "Generic (PLEG): container finished" podID="58d12e893528ad53a994f10901a644ea" containerID="248aafb616cc9e7a645eeebc252c6d5c2dc76601e14b41b4b8c8fb10f43512f6" exitCode=0 Dec 05 10:50:12.311653 master-0 kubenswrapper[24928]: I1205 10:50:12.311604 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"58d12e893528ad53a994f10901a644ea","Type":"ContainerDied","Data":"248aafb616cc9e7a645eeebc252c6d5c2dc76601e14b41b4b8c8fb10f43512f6"} Dec 05 10:50:12.312035 master-0 kubenswrapper[24928]: I1205 10:50:12.312006 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:50:12.312035 master-0 kubenswrapper[24928]: I1205 10:50:12.312033 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:50:13.323668 master-0 kubenswrapper[24928]: I1205 10:50:13.323484 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" event={"ID":"926263c4-ec5b-41cb-9c30-0c88f636035f","Type":"ContainerStarted","Data":"298a67e002c8f14ca79c236275a78968ca8b22adc515e9191b91bc9ad0d14c5f"} Dec 05 10:50:13.324213 master-0 kubenswrapper[24928]: I1205 10:50:13.324093 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:50:13.325916 master-0 kubenswrapper[24928]: I1205 10:50:13.325841 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/1.log" Dec 05 10:50:13.326542 master-0 kubenswrapper[24928]: I1205 10:50:13.326485 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-f797b99b6-z9qcl" Dec 05 10:50:13.326933 master-0 kubenswrapper[24928]: I1205 10:50:13.326879 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/0.log" Dec 05 10:50:13.327001 master-0 kubenswrapper[24928]: I1205 10:50:13.326948 24928 generic.go:334] "Generic (PLEG): container finished" podID="e27c0798-ec1c-43cd-b81b-f77f2f11ad0f" containerID="c81ba72a2f308c68cd7463f96c3158dfa0931cd098959c3eb1d285e9ee7a63b6" exitCode=1 Dec 05 10:50:13.327051 master-0 kubenswrapper[24928]: I1205 10:50:13.326993 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" event={"ID":"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f","Type":"ContainerDied","Data":"c81ba72a2f308c68cd7463f96c3158dfa0931cd098959c3eb1d285e9ee7a63b6"} Dec 05 10:50:13.327102 master-0 kubenswrapper[24928]: I1205 10:50:13.327053 24928 scope.go:117] "RemoveContainer" containerID="4308a58237214eca1f5ba917e53d41b87b4e41bf19760a2305217ef9ccb8ed3f" Dec 05 10:50:13.327849 master-0 kubenswrapper[24928]: I1205 10:50:13.327805 24928 scope.go:117] "RemoveContainer" containerID="c81ba72a2f308c68cd7463f96c3158dfa0931cd098959c3eb1d285e9ee7a63b6" Dec 05 10:50:14.336672 master-0 kubenswrapper[24928]: I1205 10:50:14.336598 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/1.log" Dec 05 10:50:14.337328 master-0 kubenswrapper[24928]: I1205 10:50:14.337230 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" event={"ID":"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f","Type":"ContainerStarted","Data":"2d6ca63681baf97c970ab633440f4db8634a012e35383b2ea20ab25e75168165"} Dec 05 10:50:14.904102 master-0 kubenswrapper[24928]: I1205 10:50:14.904023 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:50:14.904102 master-0 kubenswrapper[24928]: I1205 10:50:14.904102 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:50:17.843321 master-0 kubenswrapper[24928]: E1205 10:50:17.843170 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:50:17.843321 master-0 kubenswrapper[24928]: E1205 10:50:17.843278 24928 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 10:50:20.861199 master-0 kubenswrapper[24928]: I1205 10:50:20.861073 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:50:20.861199 master-0 kubenswrapper[24928]: I1205 10:50:20.861184 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:50:22.910651 master-0 kubenswrapper[24928]: E1205 10:50:22.910605 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="6.4s" Dec 05 10:50:24.402895 master-0 kubenswrapper[24928]: I1205 10:50:24.402833 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-catalogd_catalogd-controller-manager-7cc89f4c4c-lth87_a4fd453c-e667-4bdc-aa9e-3d95ff707200/manager/1.log" Dec 05 10:50:24.403634 master-0 kubenswrapper[24928]: I1205 10:50:24.403601 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-catalogd_catalogd-controller-manager-7cc89f4c4c-lth87_a4fd453c-e667-4bdc-aa9e-3d95ff707200/manager/0.log" Dec 05 10:50:24.404229 master-0 kubenswrapper[24928]: I1205 10:50:24.404149 24928 generic.go:334] "Generic (PLEG): container finished" podID="a4fd453c-e667-4bdc-aa9e-3d95ff707200" containerID="eaa0a2bfed5203d2ab62ebaff09a1a3afca45eb5d5bdd915133bd4c75546ae0e" exitCode=1 Dec 05 10:50:24.404292 master-0 kubenswrapper[24928]: I1205 10:50:24.404227 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" event={"ID":"a4fd453c-e667-4bdc-aa9e-3d95ff707200","Type":"ContainerDied","Data":"eaa0a2bfed5203d2ab62ebaff09a1a3afca45eb5d5bdd915133bd4c75546ae0e"} Dec 05 10:50:24.404333 master-0 kubenswrapper[24928]: I1205 10:50:24.404291 24928 scope.go:117] "RemoveContainer" containerID="2e9155fd74fd1d13dcef2794db107599bb6fa96cc533e3303f4bbbfdbd8c0208" Dec 05 10:50:24.404856 master-0 kubenswrapper[24928]: I1205 10:50:24.404832 24928 scope.go:117] "RemoveContainer" containerID="eaa0a2bfed5203d2ab62ebaff09a1a3afca45eb5d5bdd915133bd4c75546ae0e" Dec 05 10:50:24.904352 master-0 kubenswrapper[24928]: I1205 10:50:24.904240 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:50:24.904633 master-0 kubenswrapper[24928]: I1205 10:50:24.904390 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:50:25.413177 master-0 kubenswrapper[24928]: I1205 10:50:25.413126 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-catalogd_catalogd-controller-manager-7cc89f4c4c-lth87_a4fd453c-e667-4bdc-aa9e-3d95ff707200/manager/1.log" Dec 05 10:50:25.414478 master-0 kubenswrapper[24928]: I1205 10:50:25.414395 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" event={"ID":"a4fd453c-e667-4bdc-aa9e-3d95ff707200","Type":"ContainerStarted","Data":"cfc4f6aae59f29a02b65b310707479b599c6aff70105f29989253d0830dd9833"} Dec 05 10:50:25.414736 master-0 kubenswrapper[24928]: I1205 10:50:25.414696 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:50:26.397132 master-0 kubenswrapper[24928]: I1205 10:50:26.397048 24928 status_manager.go:851] "Failed to get status for pod" podUID="3c708092-6503-4a61-9230-617f8a1e2d19" pod="openshift-console/downloads-69cd4c69bf-d9jtn" err="the server was unable to return a response in the time allotted, but may still be processing the request (get pods downloads-69cd4c69bf-d9jtn)" Dec 05 10:50:29.710796 master-0 kubenswrapper[24928]: I1205 10:50:29.710713 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-lth87" Dec 05 10:50:30.450802 master-0 kubenswrapper[24928]: I1205 10:50:30.450573 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-controller-manager-operator_cluster-cloud-controller-manager-operator-758cf9d97b-74dgz_af7078c2-ad4f-415b-ba2f-77f5145c3b3f/config-sync-controllers/0.log" Dec 05 10:50:30.451406 master-0 kubenswrapper[24928]: I1205 10:50:30.451290 24928 generic.go:334] "Generic (PLEG): container finished" podID="af7078c2-ad4f-415b-ba2f-77f5145c3b3f" containerID="6ba7074f6bb0f374dabeac1deb54c10aca607e1ffaef2825067cda94e2a5c8fd" exitCode=1 Dec 05 10:50:30.451406 master-0 kubenswrapper[24928]: I1205 10:50:30.451357 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" event={"ID":"af7078c2-ad4f-415b-ba2f-77f5145c3b3f","Type":"ContainerDied","Data":"6ba7074f6bb0f374dabeac1deb54c10aca607e1ffaef2825067cda94e2a5c8fd"} Dec 05 10:50:30.452116 master-0 kubenswrapper[24928]: I1205 10:50:30.452077 24928 scope.go:117] "RemoveContainer" containerID="6ba7074f6bb0f374dabeac1deb54c10aca607e1ffaef2825067cda94e2a5c8fd" Dec 05 10:50:30.861392 master-0 kubenswrapper[24928]: I1205 10:50:30.861308 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:50:30.861392 master-0 kubenswrapper[24928]: I1205 10:50:30.861388 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:50:31.461949 master-0 kubenswrapper[24928]: I1205 10:50:31.461892 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-controller-manager-operator_cluster-cloud-controller-manager-operator-758cf9d97b-74dgz_af7078c2-ad4f-415b-ba2f-77f5145c3b3f/config-sync-controllers/0.log" Dec 05 10:50:31.462347 master-0 kubenswrapper[24928]: I1205 10:50:31.462282 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" event={"ID":"af7078c2-ad4f-415b-ba2f-77f5145c3b3f","Type":"ContainerStarted","Data":"7818c241fe08c80768cf78ae74c92c7b8c5caca1def43b96431ef30c18063725"} Dec 05 10:50:34.903872 master-0 kubenswrapper[24928]: I1205 10:50:34.903770 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:50:34.904520 master-0 kubenswrapper[24928]: I1205 10:50:34.903874 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:50:35.491478 master-0 kubenswrapper[24928]: I1205 10:50:35.491362 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-controller-manager-operator_cluster-cloud-controller-manager-operator-758cf9d97b-74dgz_af7078c2-ad4f-415b-ba2f-77f5145c3b3f/config-sync-controllers/0.log" Dec 05 10:50:35.492329 master-0 kubenswrapper[24928]: I1205 10:50:35.492268 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-controller-manager-operator_cluster-cloud-controller-manager-operator-758cf9d97b-74dgz_af7078c2-ad4f-415b-ba2f-77f5145c3b3f/cluster-cloud-controller-manager/0.log" Dec 05 10:50:35.492500 master-0 kubenswrapper[24928]: I1205 10:50:35.492330 24928 generic.go:334] "Generic (PLEG): container finished" podID="af7078c2-ad4f-415b-ba2f-77f5145c3b3f" containerID="76bf410edb844c728d030254090fc2f3989529da434d1e83243542e0c65e53f7" exitCode=1 Dec 05 10:50:35.492500 master-0 kubenswrapper[24928]: I1205 10:50:35.492370 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" event={"ID":"af7078c2-ad4f-415b-ba2f-77f5145c3b3f","Type":"ContainerDied","Data":"76bf410edb844c728d030254090fc2f3989529da434d1e83243542e0c65e53f7"} Dec 05 10:50:35.492964 master-0 kubenswrapper[24928]: I1205 10:50:35.492919 24928 scope.go:117] "RemoveContainer" containerID="76bf410edb844c728d030254090fc2f3989529da434d1e83243542e0c65e53f7" Dec 05 10:50:36.501579 master-0 kubenswrapper[24928]: I1205 10:50:36.501490 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-controller-manager-operator_cluster-cloud-controller-manager-operator-758cf9d97b-74dgz_af7078c2-ad4f-415b-ba2f-77f5145c3b3f/config-sync-controllers/0.log" Dec 05 10:50:36.502212 master-0 kubenswrapper[24928]: I1205 10:50:36.501955 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-controller-manager-operator_cluster-cloud-controller-manager-operator-758cf9d97b-74dgz_af7078c2-ad4f-415b-ba2f-77f5145c3b3f/cluster-cloud-controller-manager/0.log" Dec 05 10:50:36.502212 master-0 kubenswrapper[24928]: I1205 10:50:36.502007 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-operator-758cf9d97b-74dgz" event={"ID":"af7078c2-ad4f-415b-ba2f-77f5145c3b3f","Type":"ContainerStarted","Data":"755f9b5e7ad3df4d16578da83563d70bb19b7550b7f26a253ad8395bc42ad5be"} Dec 05 10:50:37.903070 master-0 kubenswrapper[24928]: E1205 10:50:37.902724 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:50:27Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:50:27Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:50:27Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:50:27Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:50e368e01772dd0dc9c4f9a6cdd5a9693a224968f75dc19eafe2a416f583bdab\\\"],\\\"sizeBytes\\\":2890347099},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b\\\"],\\\"sizeBytes\\\":1631758507},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:15adb3b2133604b064893f8009a74145e4c8bb5b134d111346dcccbdd2aa9bc2\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:164fc35a19aa6cc886c8015c8ee3eba4895e76b1152cb9d795e4f3154a8533a3\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1610512706},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9014f384de5f9a0b7418d5869ad349abb9588d16bd09ed650a163c045315dbff\\\"],\\\"sizeBytes\\\":1232140918},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:610b8d322265b2c9d6b07efb2be26bf4d91e428b46412d73f5bdae0218004794\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:eafb9c83c480396c34e85d1f5f5c2623be6305031245be36455850c0398bfcc7\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1209064267},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0029526507396e493c5dce1652c41ed9c239b29e84ee579a2735fdb1aa3bce83\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:e1d263cd2113e0727021ccf27c8a671f8cfeaefbf93d60e3a918d6f60c136c30\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201604946},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:e8990432556acad31519b1a73ec32f32d27c2034cf9e5cc4db8980efc7331594\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:ebe9f523f5c211a3a0f2570331dddcd5be15b12c1fecd9b8b121f881bfaad029\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1129027903},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6\\\"],\\\"sizeBytes\\\":983705650},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\\\"],\\\"sizeBytes\\\":938303566},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:01d2e67fd74086da701c39dac5b821822351cb0151f9afe72821c05df19953ad\\\"],\\\"sizeBytes\\\":912722556},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:61664aa69b33349cc6de45e44ae6033e7f483c034ea01c0d9a8ca08a12d88e3a\\\"],\\\"sizeBytes\\\":874825223},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:631a3798b749fecc041a99929eb946618df723e15055e805ff752a1a1273481c\\\"],\\\"sizeBytes\\\":870567329},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\\\"],\\\"sizeBytes\\\":857069957},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c2431a990bcddde98829abda81950247021a2ebbabc964b1516ea046b5f1d4e\\\"],\\\"sizeBytes\\\":856659740},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b12f830c3316aa4dc061c2d00c74126282b3e2bcccc301eab00d57fff3c4c7c\\\"],\\\"sizeBytes\\\":767284906},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cb3ec61f9a932a9ad13bdeb44bcf9477a8d5f728151d7f19ed3ef7d4b02b3a82\\\"],\\\"sizeBytes\\\":682371258},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:916566bb9d0143352324233d460ad94697719c11c8c9158e3aea8f475941751f\\\"],\\\"sizeBytes\\\":677523572},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b\\\"],\\\"sizeBytes\\\":672407260},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e27a636083db9043e3e4bbdc336b5e7fb5693422246e443fd1d913e157f01d46\\\"],\\\"sizeBytes\\\":628330376},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9724d2036305cbd729e1f484c5bad89971de977fff8a6723fef1873858dd1123\\\"],\\\"sizeBytes\\\":616108962},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:df606f3b71d4376d1a2108c09f0d3dab455fc30bcb67c60e91590c105e9025bf\\\"],\\\"sizeBytes\\\":583836304},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2\\\"],\\\"sizeBytes\\\":576619763},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8\\\"],\\\"sizeBytes\\\":552673986},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737\\\"],\\\"sizeBytes\\\":551889548},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc\\\"],\\\"sizeBytes\\\":543227406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718\\\"],\\\"sizeBytes\\\":532719167},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cfde59e48cd5dee3721f34d249cb119cc3259fd857965d34f9c7ed83b0c363a1\\\"],\\\"sizeBytes\\\":532402162},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f4d4282cb53325e737ad68abbfcb70687ae04fb50353f4f0ba0ba5703b15009a\\\"],\\\"sizeBytes\\\":512838054},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\\\"],\\\"sizeBytes\\\":512452153},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660\\\"],\\\"sizeBytes\\\":509437356},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:97d26892192b552c16527bf2771e1b86528ab581a02dd9279cdf71c194830e3e\\\"],\\\"sizeBytes\\\":508042119},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df\\\"],\\\"sizeBytes\\\":507687221},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8375671da86aa527ee7e291d86971b0baa823ffc7663b5a983084456e76c0f59\\\"],\\\"sizeBytes\\\":506741476},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0c3d16a01c2d60f9b536ca815ed8dc6abdca2b78e392551dc3fb79be537a354\\\"],\\\"sizeBytes\\\":506703191},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b\\\"],\\\"sizeBytes\\\":505649178},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:86af77350cfe6fd69280157e4162aa0147873d9431c641ae4ad3e881ff768a73\\\"],\\\"sizeBytes\\\":505628211},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\\\"],\\\"sizeBytes\\\":503340749},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8139ed65c0a0a4b0f253b715c11cc52be027efe8a4774da9ccce35c78ef439da\\\"],\\\"sizeBytes\\\":503011144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4\\\"],\\\"sizeBytes\\\":502436444},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10e57ca7611f79710f05777dc6a8f31c7e04eb09da4d8d793a5acfbf0e4692d7\\\"],\\\"sizeBytes\\\":500943492},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce\\\"],\\\"sizeBytes\\\":500848684},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c\\\"],\\\"sizeBytes\\\":499798563},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96\\\"],\\\"sizeBytes\\\":499705918},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33a20002692769235e95271ab071783c57ff50681088fa1035b86af31e73cf20\\\"],\\\"sizeBytes\\\":499125567},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:75d996f6147edb88c09fd1a052099de66638590d7d03a735006244bc9e19f898\\\"],\\\"sizeBytes\\\":499082775},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3b8d91a25eeb9f02041e947adb3487da3e7ab8449d3d2ad015827e7954df7b34\\\"],\\\"sizeBytes\\\":490455952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68\\\"],\\\"sizeBytes\\\":489528665},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1\\\"],\\\"sizeBytes\\\":481559117},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2b3d313c599852b3543ee5c3a62691bd2d1bbad12c2e1c610cd71a1dec6eea32\\\"],\\\"sizeBytes\\\":481499222},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3a77aa4d03b89ea284e3467a268e5989a77a2ef63e685eb1d5c5ea5b3922b7a\\\"],\\\"sizeBytes\\\":478917802}]}}\" for node \"master-0\": Patch \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0/status?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:50:38.289667 master-0 kubenswrapper[24928]: I1205 10:50:38.289583 24928 patch_prober.go:28] interesting pod/operator-controller-controller-manager-7cbd59c7f8-dh5tt container/manager namespace/openshift-operator-controller: Liveness probe status=failure output="Get \"http://10.128.0.41:8081/healthz\": dial tcp 10.128.0.41:8081: connect: connection refused" start-of-body= Dec 05 10:50:38.289667 master-0 kubenswrapper[24928]: I1205 10:50:38.289673 24928 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" podUID="592373ae-a952-4dd3-a9bc-f9c9c19c0802" containerName="manager" probeResult="failure" output="Get \"http://10.128.0.41:8081/healthz\": dial tcp 10.128.0.41:8081: connect: connection refused" Dec 05 10:50:38.467992 master-0 kubenswrapper[24928]: I1205 10:50:38.467922 24928 patch_prober.go:28] interesting pod/operator-controller-controller-manager-7cbd59c7f8-dh5tt container/manager namespace/openshift-operator-controller: Readiness probe status=failure output="Get \"http://10.128.0.41:8081/readyz\": dial tcp 10.128.0.41:8081: connect: connection refused" start-of-body= Dec 05 10:50:38.468203 master-0 kubenswrapper[24928]: I1205 10:50:38.468006 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" podUID="592373ae-a952-4dd3-a9bc-f9c9c19c0802" containerName="manager" probeResult="failure" output="Get \"http://10.128.0.41:8081/readyz\": dial tcp 10.128.0.41:8081: connect: connection refused" Dec 05 10:50:38.522173 master-0 kubenswrapper[24928]: I1205 10:50:38.522085 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-controller_operator-controller-controller-manager-7cbd59c7f8-dh5tt_592373ae-a952-4dd3-a9bc-f9c9c19c0802/manager/1.log" Dec 05 10:50:38.523042 master-0 kubenswrapper[24928]: I1205 10:50:38.522986 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-controller_operator-controller-controller-manager-7cbd59c7f8-dh5tt_592373ae-a952-4dd3-a9bc-f9c9c19c0802/manager/0.log" Dec 05 10:50:38.523042 master-0 kubenswrapper[24928]: I1205 10:50:38.523024 24928 generic.go:334] "Generic (PLEG): container finished" podID="592373ae-a952-4dd3-a9bc-f9c9c19c0802" containerID="405cb7fe7c4f15258168a2b739de72c941100f0f87222e1596be37c3a536b301" exitCode=1 Dec 05 10:50:38.523042 master-0 kubenswrapper[24928]: I1205 10:50:38.523053 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" event={"ID":"592373ae-a952-4dd3-a9bc-f9c9c19c0802","Type":"ContainerDied","Data":"405cb7fe7c4f15258168a2b739de72c941100f0f87222e1596be37c3a536b301"} Dec 05 10:50:38.523042 master-0 kubenswrapper[24928]: I1205 10:50:38.523086 24928 scope.go:117] "RemoveContainer" containerID="665b41c4b05327dcf1d9daca6aa21544857362b43e6c7d8ef7fac95d7153f9b2" Dec 05 10:50:38.524494 master-0 kubenswrapper[24928]: I1205 10:50:38.524264 24928 scope.go:117] "RemoveContainer" containerID="405cb7fe7c4f15258168a2b739de72c941100f0f87222e1596be37c3a536b301" Dec 05 10:50:39.313198 master-0 kubenswrapper[24928]: E1205 10:50:39.313066 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": context deadline exceeded" interval="7s" Dec 05 10:50:39.534094 master-0 kubenswrapper[24928]: I1205 10:50:39.534044 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operator-controller_operator-controller-controller-manager-7cbd59c7f8-dh5tt_592373ae-a952-4dd3-a9bc-f9c9c19c0802/manager/1.log" Dec 05 10:50:39.534886 master-0 kubenswrapper[24928]: I1205 10:50:39.534771 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" event={"ID":"592373ae-a952-4dd3-a9bc-f9c9c19c0802","Type":"ContainerStarted","Data":"c1f8fd9534d917e0cb1a8930f607834f4b5fe8c51764506dca70428320347738"} Dec 05 10:50:39.535204 master-0 kubenswrapper[24928]: I1205 10:50:39.535088 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:50:40.861562 master-0 kubenswrapper[24928]: I1205 10:50:40.861489 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:50:40.862170 master-0 kubenswrapper[24928]: I1205 10:50:40.861586 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:50:41.070383 master-0 kubenswrapper[24928]: E1205 10:50:41.070201 24928 event.go:359] "Server rejected event (will not retry!)" err="Timeout: request did not complete within requested timeout - context deadline exceeded" event="&Event{ObjectMeta:{downloads-69cd4c69bf-d9jtn.187e4c0928348e59 openshift-console 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-console,Name:downloads-69cd4c69bf-d9jtn,UID:3c708092-6503-4a61-9230-617f8a1e2d19,APIVersion:v1,ResourceVersion:13063,FieldPath:spec.containers{download-server},},Reason:Pulled,Message:Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:50e368e01772dd0dc9c4f9a6cdd5a9693a224968f75dc19eafe2a416f583bdab\" in 28.157s (28.157s including waiting). Image size: 2890347099 bytes.,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:48:26.188762713 +0000 UTC m=+66.191956564,LastTimestamp:2025-12-05 10:48:26.188762713 +0000 UTC m=+66.191956564,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:50:44.573047 master-0 kubenswrapper[24928]: I1205 10:50:44.572967 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/2.log" Dec 05 10:50:44.573761 master-0 kubenswrapper[24928]: I1205 10:50:44.573716 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/1.log" Dec 05 10:50:44.573828 master-0 kubenswrapper[24928]: I1205 10:50:44.573781 24928 generic.go:334] "Generic (PLEG): container finished" podID="e27c0798-ec1c-43cd-b81b-f77f2f11ad0f" containerID="2d6ca63681baf97c970ab633440f4db8634a012e35383b2ea20ab25e75168165" exitCode=1 Dec 05 10:50:44.573867 master-0 kubenswrapper[24928]: I1205 10:50:44.573824 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" event={"ID":"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f","Type":"ContainerDied","Data":"2d6ca63681baf97c970ab633440f4db8634a012e35383b2ea20ab25e75168165"} Dec 05 10:50:44.573900 master-0 kubenswrapper[24928]: I1205 10:50:44.573875 24928 scope.go:117] "RemoveContainer" containerID="c81ba72a2f308c68cd7463f96c3158dfa0931cd098959c3eb1d285e9ee7a63b6" Dec 05 10:50:44.574829 master-0 kubenswrapper[24928]: I1205 10:50:44.574677 24928 scope.go:117] "RemoveContainer" containerID="2d6ca63681baf97c970ab633440f4db8634a012e35383b2ea20ab25e75168165" Dec 05 10:50:44.575333 master-0 kubenswrapper[24928]: E1205 10:50:44.575241 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"snapshot-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=snapshot-controller pod=csi-snapshot-controller-6b958b6f94-lgn6v_openshift-cluster-storage-operator(e27c0798-ec1c-43cd-b81b-f77f2f11ad0f)\"" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" podUID="e27c0798-ec1c-43cd-b81b-f77f2f11ad0f" Dec 05 10:50:44.904730 master-0 kubenswrapper[24928]: I1205 10:50:44.904568 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:50:44.904730 master-0 kubenswrapper[24928]: I1205 10:50:44.904689 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:50:45.584158 master-0 kubenswrapper[24928]: I1205 10:50:45.584086 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/2.log" Dec 05 10:50:46.315410 master-0 kubenswrapper[24928]: E1205 10:50:46.315330 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Timeout: request did not complete within requested timeout - context deadline exceeded" pod="openshift-etcd/etcd-master-0" Dec 05 10:50:46.594280 master-0 kubenswrapper[24928]: I1205 10:50:46.594146 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-7df95c79b5-qnq6t_24444474-aa9a-4a0f-8b4d-90f0009e0dc7/control-plane-machine-set-operator/0.log" Dec 05 10:50:46.594280 master-0 kubenswrapper[24928]: I1205 10:50:46.594215 24928 generic.go:334] "Generic (PLEG): container finished" podID="24444474-aa9a-4a0f-8b4d-90f0009e0dc7" containerID="f67f402cd7e32dc638b6b2ed5376d39de0b7eae17574824302e014030169c159" exitCode=1 Dec 05 10:50:46.595058 master-0 kubenswrapper[24928]: I1205 10:50:46.594281 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" event={"ID":"24444474-aa9a-4a0f-8b4d-90f0009e0dc7","Type":"ContainerDied","Data":"f67f402cd7e32dc638b6b2ed5376d39de0b7eae17574824302e014030169c159"} Dec 05 10:50:46.595058 master-0 kubenswrapper[24928]: I1205 10:50:46.594879 24928 scope.go:117] "RemoveContainer" containerID="f67f402cd7e32dc638b6b2ed5376d39de0b7eae17574824302e014030169c159" Dec 05 10:50:46.597289 master-0 kubenswrapper[24928]: I1205 10:50:46.596955 24928 generic.go:334] "Generic (PLEG): container finished" podID="b760849c-8d83-47da-8677-68445c143bef" containerID="0423f72dfd535a3ee285bcc8d731216360e562171918005d692bba035e2bd6b5" exitCode=0 Dec 05 10:50:46.597289 master-0 kubenswrapper[24928]: I1205 10:50:46.597022 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" event={"ID":"b760849c-8d83-47da-8677-68445c143bef","Type":"ContainerDied","Data":"0423f72dfd535a3ee285bcc8d731216360e562171918005d692bba035e2bd6b5"} Dec 05 10:50:46.597289 master-0 kubenswrapper[24928]: I1205 10:50:46.597139 24928 scope.go:117] "RemoveContainer" containerID="d5454b7ce1bc247671ab46448edd5a6f0a198a1673e48f822fea3525f6db868a" Dec 05 10:50:46.597535 master-0 kubenswrapper[24928]: I1205 10:50:46.597518 24928 scope.go:117] "RemoveContainer" containerID="0423f72dfd535a3ee285bcc8d731216360e562171918005d692bba035e2bd6b5" Dec 05 10:50:47.903876 master-0 kubenswrapper[24928]: E1205 10:50:47.903760 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:50:48.469722 master-0 kubenswrapper[24928]: I1205 10:50:48.469658 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-controller/operator-controller-controller-manager-7cbd59c7f8-dh5tt" Dec 05 10:50:48.617377 master-0 kubenswrapper[24928]: I1205 10:50:48.617295 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-5df5548d54-gr5gp" event={"ID":"b760849c-8d83-47da-8677-68445c143bef","Type":"ContainerStarted","Data":"f08ae502ea535befd4d6f8de6788ece1116e919453caf94407c14141bf7a9baa"} Dec 05 10:50:48.619852 master-0 kubenswrapper[24928]: I1205 10:50:48.619816 24928 generic.go:334] "Generic (PLEG): container finished" podID="58d12e893528ad53a994f10901a644ea" containerID="12cb4e737c9c66d2fe3cc4bd86482f28a548a5b5cc2c79adfbd5d604b334960f" exitCode=0 Dec 05 10:50:48.620152 master-0 kubenswrapper[24928]: I1205 10:50:48.619881 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"58d12e893528ad53a994f10901a644ea","Type":"ContainerDied","Data":"12cb4e737c9c66d2fe3cc4bd86482f28a548a5b5cc2c79adfbd5d604b334960f"} Dec 05 10:50:48.620243 master-0 kubenswrapper[24928]: I1205 10:50:48.620219 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:50:48.620311 master-0 kubenswrapper[24928]: I1205 10:50:48.620251 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:50:48.621970 master-0 kubenswrapper[24928]: I1205 10:50:48.621927 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-7df95c79b5-qnq6t_24444474-aa9a-4a0f-8b4d-90f0009e0dc7/control-plane-machine-set-operator/0.log" Dec 05 10:50:48.622066 master-0 kubenswrapper[24928]: I1205 10:50:48.621984 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-qnq6t" event={"ID":"24444474-aa9a-4a0f-8b4d-90f0009e0dc7","Type":"ContainerStarted","Data":"6a30faaa5f7ea37d220e63813d47c2cbc3c0f15fc0f64be4233cecad1e4c6a98"} Dec 05 10:50:50.897857 master-0 kubenswrapper[24928]: I1205 10:50:50.861610 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:50:50.897857 master-0 kubenswrapper[24928]: I1205 10:50:50.861679 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:50:52.650439 master-0 kubenswrapper[24928]: I1205 10:50:52.650366 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-baremetal-operator-78f758c7b9-6t2gm_48bd1d86-a6f2-439f-ab04-6a9a442bec42/cluster-baremetal-operator/0.log" Dec 05 10:50:52.651066 master-0 kubenswrapper[24928]: I1205 10:50:52.650453 24928 generic.go:334] "Generic (PLEG): container finished" podID="48bd1d86-a6f2-439f-ab04-6a9a442bec42" containerID="3a9cca203c845359b6a28fb821277d5c1db034d15078e96e3ff87b488109d4de" exitCode=1 Dec 05 10:50:52.651066 master-0 kubenswrapper[24928]: I1205 10:50:52.650511 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" event={"ID":"48bd1d86-a6f2-439f-ab04-6a9a442bec42","Type":"ContainerDied","Data":"3a9cca203c845359b6a28fb821277d5c1db034d15078e96e3ff87b488109d4de"} Dec 05 10:50:52.651066 master-0 kubenswrapper[24928]: I1205 10:50:52.651004 24928 scope.go:117] "RemoveContainer" containerID="3a9cca203c845359b6a28fb821277d5c1db034d15078e96e3ff87b488109d4de" Dec 05 10:50:52.652354 master-0 kubenswrapper[24928]: I1205 10:50:52.652295 24928 generic.go:334] "Generic (PLEG): container finished" podID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerID="c3968a3a8ccd5212aba3d71d18db810366b8e5989f94f317e2256250850f0562" exitCode=0 Dec 05 10:50:52.652413 master-0 kubenswrapper[24928]: I1205 10:50:52.652362 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" event={"ID":"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0","Type":"ContainerDied","Data":"c3968a3a8ccd5212aba3d71d18db810366b8e5989f94f317e2256250850f0562"} Dec 05 10:50:52.652491 master-0 kubenswrapper[24928]: I1205 10:50:52.652411 24928 scope.go:117] "RemoveContainer" containerID="8aa1ac52d0f6c8b159f0da42738bae856c9e11f506fdfed545be3d79a9c5ba1d" Dec 05 10:50:52.652760 master-0 kubenswrapper[24928]: I1205 10:50:52.652739 24928 scope.go:117] "RemoveContainer" containerID="c3968a3a8ccd5212aba3d71d18db810366b8e5989f94f317e2256250850f0562" Dec 05 10:50:53.669115 master-0 kubenswrapper[24928]: I1205 10:50:53.669046 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-baremetal-operator-78f758c7b9-6t2gm_48bd1d86-a6f2-439f-ab04-6a9a442bec42/cluster-baremetal-operator/0.log" Dec 05 10:50:53.670093 master-0 kubenswrapper[24928]: I1205 10:50:53.669142 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" event={"ID":"48bd1d86-a6f2-439f-ab04-6a9a442bec42","Type":"ContainerStarted","Data":"10fa0576349384ec32341165d84160e0c9b5b5d6f5044617a559a0d068966dc5"} Dec 05 10:50:53.671123 master-0 kubenswrapper[24928]: I1205 10:50:53.671074 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" event={"ID":"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0","Type":"ContainerStarted","Data":"66ed166decce5e8377b96bf40238cb4562c34424e0d591957563972e8e5b494f"} Dec 05 10:50:53.671398 master-0 kubenswrapper[24928]: I1205 10:50:53.671319 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:50:53.674966 master-0 kubenswrapper[24928]: I1205 10:50:53.674903 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:50:54.904458 master-0 kubenswrapper[24928]: I1205 10:50:54.904323 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:50:54.904458 master-0 kubenswrapper[24928]: I1205 10:50:54.904404 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:50:55.204803 master-0 kubenswrapper[24928]: I1205 10:50:55.204646 24928 scope.go:117] "RemoveContainer" containerID="2d6ca63681baf97c970ab633440f4db8634a012e35383b2ea20ab25e75168165" Dec 05 10:50:55.687483 master-0 kubenswrapper[24928]: I1205 10:50:55.687396 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/2.log" Dec 05 10:50:55.687759 master-0 kubenswrapper[24928]: I1205 10:50:55.687571 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" event={"ID":"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f","Type":"ContainerStarted","Data":"d31f42cb2df60e0cf78918872183788f5baa6dfb0312a552f84381b8462fd7e0"} Dec 05 10:50:55.690739 master-0 kubenswrapper[24928]: I1205 10:50:55.690689 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/0.log" Dec 05 10:50:55.690837 master-0 kubenswrapper[24928]: I1205 10:50:55.690760 24928 generic.go:334] "Generic (PLEG): container finished" podID="5219435a07a0220d41da97c4fb70abb1" containerID="f7c86a3fbacaa184c777e4b5c80c29c00ccd9777b6aeed8a46d62880a41cc869" exitCode=0 Dec 05 10:50:55.690837 master-0 kubenswrapper[24928]: I1205 10:50:55.690792 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerDied","Data":"f7c86a3fbacaa184c777e4b5c80c29c00ccd9777b6aeed8a46d62880a41cc869"} Dec 05 10:50:55.691400 master-0 kubenswrapper[24928]: I1205 10:50:55.691369 24928 scope.go:117] "RemoveContainer" containerID="f7c86a3fbacaa184c777e4b5c80c29c00ccd9777b6aeed8a46d62880a41cc869" Dec 05 10:50:56.315085 master-0 kubenswrapper[24928]: E1205 10:50:56.314886 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="7s" Dec 05 10:50:56.702763 master-0 kubenswrapper[24928]: I1205 10:50:56.702583 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/0.log" Dec 05 10:50:56.702763 master-0 kubenswrapper[24928]: I1205 10:50:56.702652 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"50e85593f8bade43c99e607efe092eba239c2c489a561c3d4ee019a9a7b227cb"} Dec 05 10:50:57.904757 master-0 kubenswrapper[24928]: E1205 10:50:57.904631 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:51:00.861247 master-0 kubenswrapper[24928]: I1205 10:51:00.861182 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:51:00.861996 master-0 kubenswrapper[24928]: I1205 10:51:00.861964 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:51:01.742159 master-0 kubenswrapper[24928]: I1205 10:51:01.742128 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/0.log" Dec 05 10:51:01.742495 master-0 kubenswrapper[24928]: I1205 10:51:01.742471 24928 generic.go:334] "Generic (PLEG): container finished" podID="5219435a07a0220d41da97c4fb70abb1" containerID="00db1aab947f0a6caa3edb56d388eb498e266a62880572bf52a5608eb626a365" exitCode=0 Dec 05 10:51:01.742609 master-0 kubenswrapper[24928]: I1205 10:51:01.742575 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerDied","Data":"00db1aab947f0a6caa3edb56d388eb498e266a62880572bf52a5608eb626a365"} Dec 05 10:51:01.743172 master-0 kubenswrapper[24928]: I1205 10:51:01.743152 24928 scope.go:117] "RemoveContainer" containerID="00db1aab947f0a6caa3edb56d388eb498e266a62880572bf52a5608eb626a365" Dec 05 10:51:02.752990 master-0 kubenswrapper[24928]: I1205 10:51:02.752786 24928 generic.go:334] "Generic (PLEG): container finished" podID="bdd20c68f1ccfd1240fb273af0638d01" containerID="29ac63e75838b26e5ffeb47eea93679b6f563ae1d361a9d774732d46e7d9c13d" exitCode=0 Dec 05 10:51:02.752990 master-0 kubenswrapper[24928]: I1205 10:51:02.752882 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerDied","Data":"29ac63e75838b26e5ffeb47eea93679b6f563ae1d361a9d774732d46e7d9c13d"} Dec 05 10:51:02.753900 master-0 kubenswrapper[24928]: I1205 10:51:02.753501 24928 scope.go:117] "RemoveContainer" containerID="29ac63e75838b26e5ffeb47eea93679b6f563ae1d361a9d774732d46e7d9c13d" Dec 05 10:51:02.757514 master-0 kubenswrapper[24928]: I1205 10:51:02.756651 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/0.log" Dec 05 10:51:02.757514 master-0 kubenswrapper[24928]: I1205 10:51:02.756724 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"b81e8b68cc34c8fa308242ed9c31962b7b0b1e3af80c9d0154c5d9ae32a5982d"} Dec 05 10:51:03.242105 master-0 kubenswrapper[24928]: I1205 10:51:03.241996 24928 patch_prober.go:28] interesting pod/openshift-kube-scheduler-master-0 container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.32.10:10259/healthz\": dial tcp 192.168.32.10:10259: connect: connection refused" start-of-body= Dec 05 10:51:03.242400 master-0 kubenswrapper[24928]: I1205 10:51:03.242118 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" podUID="bdd20c68f1ccfd1240fb273af0638d01" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.32.10:10259/healthz\": dial tcp 192.168.32.10:10259: connect: connection refused" Dec 05 10:51:03.766546 master-0 kubenswrapper[24928]: I1205 10:51:03.766485 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-machine-approver_machine-approver-74d9cbffbc-9jbnk_62f174ff-9f3c-4051-a34a-e55a59207171/machine-approver-controller/0.log" Dec 05 10:51:03.767143 master-0 kubenswrapper[24928]: I1205 10:51:03.766815 24928 generic.go:334] "Generic (PLEG): container finished" podID="62f174ff-9f3c-4051-a34a-e55a59207171" containerID="f445628393287d820027f18b6a0fdccb35344ae04fb12bc19d9b3cf5ae9e9d49" exitCode=255 Dec 05 10:51:03.767143 master-0 kubenswrapper[24928]: I1205 10:51:03.766889 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" event={"ID":"62f174ff-9f3c-4051-a34a-e55a59207171","Type":"ContainerDied","Data":"f445628393287d820027f18b6a0fdccb35344ae04fb12bc19d9b3cf5ae9e9d49"} Dec 05 10:51:03.767660 master-0 kubenswrapper[24928]: I1205 10:51:03.767616 24928 scope.go:117] "RemoveContainer" containerID="f445628393287d820027f18b6a0fdccb35344ae04fb12bc19d9b3cf5ae9e9d49" Dec 05 10:51:03.770127 master-0 kubenswrapper[24928]: I1205 10:51:03.770090 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler_openshift-kube-scheduler-master-0_bdd20c68f1ccfd1240fb273af0638d01/kube-scheduler/0.log" Dec 05 10:51:03.770632 master-0 kubenswrapper[24928]: I1205 10:51:03.770576 24928 generic.go:334] "Generic (PLEG): container finished" podID="bdd20c68f1ccfd1240fb273af0638d01" containerID="8c491ff443ec7786bd22bb8e647eeff7f1b6e812a627f82eec65955c76037686" exitCode=1 Dec 05 10:51:03.770697 master-0 kubenswrapper[24928]: I1205 10:51:03.770637 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerDied","Data":"8c491ff443ec7786bd22bb8e647eeff7f1b6e812a627f82eec65955c76037686"} Dec 05 10:51:03.770697 master-0 kubenswrapper[24928]: I1205 10:51:03.770674 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerStarted","Data":"380db68fca89bbaa4ae24892ce5fc6be0a791c88eacb32a005087d29df2fad21"} Dec 05 10:51:03.771042 master-0 kubenswrapper[24928]: I1205 10:51:03.771019 24928 scope.go:117] "RemoveContainer" containerID="8c491ff443ec7786bd22bb8e647eeff7f1b6e812a627f82eec65955c76037686" Dec 05 10:51:03.810747 master-0 kubenswrapper[24928]: I1205 10:51:03.810701 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:51:04.778122 master-0 kubenswrapper[24928]: I1205 10:51:04.778055 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-machine-approver_machine-approver-74d9cbffbc-9jbnk_62f174ff-9f3c-4051-a34a-e55a59207171/machine-approver-controller/0.log" Dec 05 10:51:04.778729 master-0 kubenswrapper[24928]: I1205 10:51:04.778628 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-74d9cbffbc-9jbnk" event={"ID":"62f174ff-9f3c-4051-a34a-e55a59207171","Type":"ContainerStarted","Data":"c92b35922c98bcd83f2fdc295b1ace264d310c6312ed6368e80437f72a9968f8"} Dec 05 10:51:04.781907 master-0 kubenswrapper[24928]: I1205 10:51:04.781862 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-scheduler_openshift-kube-scheduler-master-0_bdd20c68f1ccfd1240fb273af0638d01/kube-scheduler/0.log" Dec 05 10:51:04.782336 master-0 kubenswrapper[24928]: I1205 10:51:04.782280 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" event={"ID":"bdd20c68f1ccfd1240fb273af0638d01","Type":"ContainerStarted","Data":"5225f8be214ce3fca7a48c73a9a3cecc61a716f17f6ceccd3e4a92f55ced03e4"} Dec 05 10:51:04.782649 master-0 kubenswrapper[24928]: I1205 10:51:04.782575 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:51:04.903480 master-0 kubenswrapper[24928]: I1205 10:51:04.903397 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:51:04.903480 master-0 kubenswrapper[24928]: I1205 10:51:04.903481 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:51:06.811315 master-0 kubenswrapper[24928]: I1205 10:51:06.811226 24928 patch_prober.go:28] interesting pod/kube-controller-manager-master-0 container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 10:51:06.812009 master-0 kubenswrapper[24928]: I1205 10:51:06.811329 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 10:51:07.905983 master-0 kubenswrapper[24928]: E1205 10:51:07.905854 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:51:09.580975 master-0 kubenswrapper[24928]: I1205 10:51:09.580891 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:51:10.861661 master-0 kubenswrapper[24928]: I1205 10:51:10.861598 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:51:10.862187 master-0 kubenswrapper[24928]: I1205 10:51:10.861665 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:51:13.316571 master-0 kubenswrapper[24928]: E1205 10:51:13.316451 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="7s" Dec 05 10:51:14.904551 master-0 kubenswrapper[24928]: I1205 10:51:14.904470 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:51:14.905243 master-0 kubenswrapper[24928]: I1205 10:51:14.904554 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:51:15.073329 master-0 kubenswrapper[24928]: E1205 10:51:15.073096 24928 event.go:359] "Server rejected event (will not retry!)" err="Timeout: request did not complete within requested timeout - context deadline exceeded" event="&Event{ObjectMeta:{downloads-69cd4c69bf-d9jtn.187e4c09313d8b1c openshift-console 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-console,Name:downloads-69cd4c69bf-d9jtn,UID:3c708092-6503-4a61-9230-617f8a1e2d19,APIVersion:v1,ResourceVersion:13063,FieldPath:spec.containers{download-server},},Reason:Created,Message:Created container: download-server,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:48:26.340346652 +0000 UTC m=+66.343540503,LastTimestamp:2025-12-05 10:48:26.340346652 +0000 UTC m=+66.343540503,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:51:16.810632 master-0 kubenswrapper[24928]: I1205 10:51:16.810540 24928 patch_prober.go:28] interesting pod/kube-controller-manager-master-0 container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 10:51:16.810632 master-0 kubenswrapper[24928]: I1205 10:51:16.810619 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 10:51:17.906515 master-0 kubenswrapper[24928]: E1205 10:51:17.906342 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:51:17.906515 master-0 kubenswrapper[24928]: E1205 10:51:17.906393 24928 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 10:51:20.862145 master-0 kubenswrapper[24928]: I1205 10:51:20.862057 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:51:20.862145 master-0 kubenswrapper[24928]: I1205 10:51:20.862136 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:51:22.622808 master-0 kubenswrapper[24928]: E1205 10:51:22.622738 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Timeout: request did not complete within requested timeout - context deadline exceeded" pod="openshift-etcd/etcd-master-0" Dec 05 10:51:23.905318 master-0 kubenswrapper[24928]: I1205 10:51:23.905244 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"58d12e893528ad53a994f10901a644ea","Type":"ContainerStarted","Data":"349ce0f25655b8eafee0223909bcba5b7fcea0fd278ec922bd70f16532f5a83c"} Dec 05 10:51:23.905318 master-0 kubenswrapper[24928]: I1205 10:51:23.905312 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"58d12e893528ad53a994f10901a644ea","Type":"ContainerStarted","Data":"65e7dc775f9b903f864d03460f97b58c8c0e7629adb4c1a2a38f2ad73d7c39ab"} Dec 05 10:51:23.905318 master-0 kubenswrapper[24928]: I1205 10:51:23.905327 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"58d12e893528ad53a994f10901a644ea","Type":"ContainerStarted","Data":"5fb61e8b9729625ae43ae87d4d6128e709c2bd2a1e2e471d1e03d8c13cc548da"} Dec 05 10:51:24.904560 master-0 kubenswrapper[24928]: I1205 10:51:24.904391 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:51:24.904560 master-0 kubenswrapper[24928]: I1205 10:51:24.904533 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:51:24.919399 master-0 kubenswrapper[24928]: I1205 10:51:24.919345 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"58d12e893528ad53a994f10901a644ea","Type":"ContainerStarted","Data":"7af113abd305fa6b983b07383f4fbdc4ddb7dda77e1cb7229fe9aa7569f577a8"} Dec 05 10:51:25.932107 master-0 kubenswrapper[24928]: I1205 10:51:25.931905 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-master-0" event={"ID":"58d12e893528ad53a994f10901a644ea","Type":"ContainerStarted","Data":"0196da7ea350db2032f72563c90e8277b2f9a6bc34a822a931025e1ee2729cf8"} Dec 05 10:51:25.932598 master-0 kubenswrapper[24928]: I1205 10:51:25.932128 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:51:25.932598 master-0 kubenswrapper[24928]: I1205 10:51:25.932167 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:51:25.933460 master-0 kubenswrapper[24928]: I1205 10:51:25.933401 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/3.log" Dec 05 10:51:25.933867 master-0 kubenswrapper[24928]: I1205 10:51:25.933827 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/2.log" Dec 05 10:51:25.933937 master-0 kubenswrapper[24928]: I1205 10:51:25.933878 24928 generic.go:334] "Generic (PLEG): container finished" podID="e27c0798-ec1c-43cd-b81b-f77f2f11ad0f" containerID="d31f42cb2df60e0cf78918872183788f5baa6dfb0312a552f84381b8462fd7e0" exitCode=1 Dec 05 10:51:25.933995 master-0 kubenswrapper[24928]: I1205 10:51:25.933928 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" event={"ID":"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f","Type":"ContainerDied","Data":"d31f42cb2df60e0cf78918872183788f5baa6dfb0312a552f84381b8462fd7e0"} Dec 05 10:51:25.934083 master-0 kubenswrapper[24928]: I1205 10:51:25.934042 24928 scope.go:117] "RemoveContainer" containerID="2d6ca63681baf97c970ab633440f4db8634a012e35383b2ea20ab25e75168165" Dec 05 10:51:25.934540 master-0 kubenswrapper[24928]: I1205 10:51:25.934508 24928 scope.go:117] "RemoveContainer" containerID="d31f42cb2df60e0cf78918872183788f5baa6dfb0312a552f84381b8462fd7e0" Dec 05 10:51:25.934772 master-0 kubenswrapper[24928]: E1205 10:51:25.934740 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"snapshot-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=snapshot-controller pod=csi-snapshot-controller-6b958b6f94-lgn6v_openshift-cluster-storage-operator(e27c0798-ec1c-43cd-b81b-f77f2f11ad0f)\"" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" podUID="e27c0798-ec1c-43cd-b81b-f77f2f11ad0f" Dec 05 10:51:26.232012 master-0 kubenswrapper[24928]: I1205 10:51:26.231872 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-master-0" Dec 05 10:51:26.232012 master-0 kubenswrapper[24928]: I1205 10:51:26.231933 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-master-0" Dec 05 10:51:26.399514 master-0 kubenswrapper[24928]: I1205 10:51:26.399397 24928 status_manager.go:851] "Failed to get status for pod" podUID="58d12e893528ad53a994f10901a644ea" pod="openshift-etcd/etcd-master-0" err="the server was unable to return a response in the time allotted, but may still be processing the request (get pods etcd-master-0)" Dec 05 10:51:26.809972 master-0 kubenswrapper[24928]: I1205 10:51:26.809869 24928 patch_prober.go:28] interesting pod/kube-controller-manager-master-0 container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 10:51:26.809972 master-0 kubenswrapper[24928]: I1205 10:51:26.809969 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 10:51:26.810329 master-0 kubenswrapper[24928]: I1205 10:51:26.810030 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:51:26.810754 master-0 kubenswrapper[24928]: I1205 10:51:26.810719 24928 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"b81e8b68cc34c8fa308242ed9c31962b7b0b1e3af80c9d0154c5d9ae32a5982d"} pod="openshift-kube-controller-manager/kube-controller-manager-master-0" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Dec 05 10:51:26.810850 master-0 kubenswrapper[24928]: I1205 10:51:26.810825 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" containerID="cri-o://b81e8b68cc34c8fa308242ed9c31962b7b0b1e3af80c9d0154c5d9ae32a5982d" gracePeriod=30 Dec 05 10:51:26.943255 master-0 kubenswrapper[24928]: I1205 10:51:26.943222 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/3.log" Dec 05 10:51:27.957251 master-0 kubenswrapper[24928]: I1205 10:51:27.957203 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/cluster-policy-controller/1.log" Dec 05 10:51:27.959415 master-0 kubenswrapper[24928]: I1205 10:51:27.959374 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/0.log" Dec 05 10:51:27.959415 master-0 kubenswrapper[24928]: I1205 10:51:27.959418 24928 generic.go:334] "Generic (PLEG): container finished" podID="5219435a07a0220d41da97c4fb70abb1" containerID="b81e8b68cc34c8fa308242ed9c31962b7b0b1e3af80c9d0154c5d9ae32a5982d" exitCode=255 Dec 05 10:51:27.959613 master-0 kubenswrapper[24928]: I1205 10:51:27.959460 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerDied","Data":"b81e8b68cc34c8fa308242ed9c31962b7b0b1e3af80c9d0154c5d9ae32a5982d"} Dec 05 10:51:27.959613 master-0 kubenswrapper[24928]: I1205 10:51:27.959494 24928 scope.go:117] "RemoveContainer" containerID="00db1aab947f0a6caa3edb56d388eb498e266a62880572bf52a5608eb626a365" Dec 05 10:51:28.968728 master-0 kubenswrapper[24928]: I1205 10:51:28.968684 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/cluster-policy-controller/1.log" Dec 05 10:51:28.970574 master-0 kubenswrapper[24928]: I1205 10:51:28.970554 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/0.log" Dec 05 10:51:28.970645 master-0 kubenswrapper[24928]: I1205 10:51:28.970617 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"b27b1cff74cf1fa7645c9d670ba8742571627f768b820d46bb086bfbeb3ea29b"} Dec 05 10:51:29.580879 master-0 kubenswrapper[24928]: I1205 10:51:29.580796 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:51:30.320153 master-0 kubenswrapper[24928]: E1205 10:51:30.320070 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io master-0)" interval="7s" Dec 05 10:51:30.861547 master-0 kubenswrapper[24928]: I1205 10:51:30.861477 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:51:30.861784 master-0 kubenswrapper[24928]: I1205 10:51:30.861573 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:51:33.810141 master-0 kubenswrapper[24928]: I1205 10:51:33.810068 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:51:33.814311 master-0 kubenswrapper[24928]: I1205 10:51:33.814273 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:51:34.903777 master-0 kubenswrapper[24928]: I1205 10:51:34.903714 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:51:34.903777 master-0 kubenswrapper[24928]: I1205 10:51:34.903778 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:51:36.259130 master-0 kubenswrapper[24928]: I1205 10:51:36.259081 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-master-0" Dec 05 10:51:38.191468 master-0 kubenswrapper[24928]: E1205 10:51:38.191274 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:51:28Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:51:28Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:51:28Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:51:28Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:50e368e01772dd0dc9c4f9a6cdd5a9693a224968f75dc19eafe2a416f583bdab\\\"],\\\"sizeBytes\\\":2890347099},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b\\\"],\\\"sizeBytes\\\":1631758507},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:15adb3b2133604b064893f8009a74145e4c8bb5b134d111346dcccbdd2aa9bc2\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:164fc35a19aa6cc886c8015c8ee3eba4895e76b1152cb9d795e4f3154a8533a3\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1610512706},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9014f384de5f9a0b7418d5869ad349abb9588d16bd09ed650a163c045315dbff\\\"],\\\"sizeBytes\\\":1232140918},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:610b8d322265b2c9d6b07efb2be26bf4d91e428b46412d73f5bdae0218004794\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:eafb9c83c480396c34e85d1f5f5c2623be6305031245be36455850c0398bfcc7\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1209064267},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0029526507396e493c5dce1652c41ed9c239b29e84ee579a2735fdb1aa3bce83\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:e1d263cd2113e0727021ccf27c8a671f8cfeaefbf93d60e3a918d6f60c136c30\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201604946},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:e8990432556acad31519b1a73ec32f32d27c2034cf9e5cc4db8980efc7331594\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:ebe9f523f5c211a3a0f2570331dddcd5be15b12c1fecd9b8b121f881bfaad029\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1129027903},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6\\\"],\\\"sizeBytes\\\":983705650},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\\\"],\\\"sizeBytes\\\":938303566},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:01d2e67fd74086da701c39dac5b821822351cb0151f9afe72821c05df19953ad\\\"],\\\"sizeBytes\\\":912722556},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:61664aa69b33349cc6de45e44ae6033e7f483c034ea01c0d9a8ca08a12d88e3a\\\"],\\\"sizeBytes\\\":874825223},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:631a3798b749fecc041a99929eb946618df723e15055e805ff752a1a1273481c\\\"],\\\"sizeBytes\\\":870567329},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\\\"],\\\"sizeBytes\\\":857069957},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c2431a990bcddde98829abda81950247021a2ebbabc964b1516ea046b5f1d4e\\\"],\\\"sizeBytes\\\":856659740},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b12f830c3316aa4dc061c2d00c74126282b3e2bcccc301eab00d57fff3c4c7c\\\"],\\\"sizeBytes\\\":767284906},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cb3ec61f9a932a9ad13bdeb44bcf9477a8d5f728151d7f19ed3ef7d4b02b3a82\\\"],\\\"sizeBytes\\\":682371258},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:916566bb9d0143352324233d460ad94697719c11c8c9158e3aea8f475941751f\\\"],\\\"sizeBytes\\\":677523572},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b\\\"],\\\"sizeBytes\\\":672407260},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e27a636083db9043e3e4bbdc336b5e7fb5693422246e443fd1d913e157f01d46\\\"],\\\"sizeBytes\\\":628330376},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9724d2036305cbd729e1f484c5bad89971de977fff8a6723fef1873858dd1123\\\"],\\\"sizeBytes\\\":616108962},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:df606f3b71d4376d1a2108c09f0d3dab455fc30bcb67c60e91590c105e9025bf\\\"],\\\"sizeBytes\\\":583836304},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2\\\"],\\\"sizeBytes\\\":576619763},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8\\\"],\\\"sizeBytes\\\":552673986},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737\\\"],\\\"sizeBytes\\\":551889548},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc\\\"],\\\"sizeBytes\\\":543227406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718\\\"],\\\"sizeBytes\\\":532719167},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cfde59e48cd5dee3721f34d249cb119cc3259fd857965d34f9c7ed83b0c363a1\\\"],\\\"sizeBytes\\\":532402162},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f4d4282cb53325e737ad68abbfcb70687ae04fb50353f4f0ba0ba5703b15009a\\\"],\\\"sizeBytes\\\":512838054},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\\\"],\\\"sizeBytes\\\":512452153},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660\\\"],\\\"sizeBytes\\\":509437356},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:97d26892192b552c16527bf2771e1b86528ab581a02dd9279cdf71c194830e3e\\\"],\\\"sizeBytes\\\":508042119},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df\\\"],\\\"sizeBytes\\\":507687221},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8375671da86aa527ee7e291d86971b0baa823ffc7663b5a983084456e76c0f59\\\"],\\\"sizeBytes\\\":506741476},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0c3d16a01c2d60f9b536ca815ed8dc6abdca2b78e392551dc3fb79be537a354\\\"],\\\"sizeBytes\\\":506703191},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b\\\"],\\\"sizeBytes\\\":505649178},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:86af77350cfe6fd69280157e4162aa0147873d9431c641ae4ad3e881ff768a73\\\"],\\\"sizeBytes\\\":505628211},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\\\"],\\\"sizeBytes\\\":503340749},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8139ed65c0a0a4b0f253b715c11cc52be027efe8a4774da9ccce35c78ef439da\\\"],\\\"sizeBytes\\\":503011144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4\\\"],\\\"sizeBytes\\\":502436444},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10e57ca7611f79710f05777dc6a8f31c7e04eb09da4d8d793a5acfbf0e4692d7\\\"],\\\"sizeBytes\\\":500943492},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce\\\"],\\\"sizeBytes\\\":500848684},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c\\\"],\\\"sizeBytes\\\":499798563},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96\\\"],\\\"sizeBytes\\\":499705918},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33a20002692769235e95271ab071783c57ff50681088fa1035b86af31e73cf20\\\"],\\\"sizeBytes\\\":499125567},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:75d996f6147edb88c09fd1a052099de66638590d7d03a735006244bc9e19f898\\\"],\\\"sizeBytes\\\":499082775},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3b8d91a25eeb9f02041e947adb3487da3e7ab8449d3d2ad015827e7954df7b34\\\"],\\\"sizeBytes\\\":490455952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68\\\"],\\\"sizeBytes\\\":489528665},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1\\\"],\\\"sizeBytes\\\":481559117},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2b3d313c599852b3543ee5c3a62691bd2d1bbad12c2e1c610cd71a1dec6eea32\\\"],\\\"sizeBytes\\\":481499222},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3a77aa4d03b89ea284e3467a268e5989a77a2ef63e685eb1d5c5ea5b3922b7a\\\"],\\\"sizeBytes\\\":478917802}]}}\" for node \"master-0\": Patch \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0/status?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:51:39.586212 master-0 kubenswrapper[24928]: I1205 10:51:39.586116 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:51:40.204145 master-0 kubenswrapper[24928]: I1205 10:51:40.204066 24928 scope.go:117] "RemoveContainer" containerID="d31f42cb2df60e0cf78918872183788f5baa6dfb0312a552f84381b8462fd7e0" Dec 05 10:51:40.204585 master-0 kubenswrapper[24928]: E1205 10:51:40.204460 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"snapshot-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=snapshot-controller pod=csi-snapshot-controller-6b958b6f94-lgn6v_openshift-cluster-storage-operator(e27c0798-ec1c-43cd-b81b-f77f2f11ad0f)\"" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" podUID="e27c0798-ec1c-43cd-b81b-f77f2f11ad0f" Dec 05 10:51:40.860870 master-0 kubenswrapper[24928]: I1205 10:51:40.860797 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:51:40.860870 master-0 kubenswrapper[24928]: I1205 10:51:40.860845 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:51:41.255350 master-0 kubenswrapper[24928]: I1205 10:51:41.255265 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-master-0" Dec 05 10:51:44.904282 master-0 kubenswrapper[24928]: I1205 10:51:44.904151 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:51:44.904282 master-0 kubenswrapper[24928]: I1205 10:51:44.904256 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:51:49.076253 master-0 kubenswrapper[24928]: E1205 10:51:49.075917 24928 event.go:359] "Server rejected event (will not retry!)" err="Timeout: request did not complete within requested timeout - context deadline exceeded" event="&Event{ObjectMeta:{downloads-69cd4c69bf-d9jtn.187e4c0931cbc0ea openshift-console 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-console,Name:downloads-69cd4c69bf-d9jtn,UID:3c708092-6503-4a61-9230-617f8a1e2d19,APIVersion:v1,ResourceVersion:13063,FieldPath:spec.containers{download-server},},Reason:Started,Message:Started container download-server,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:48:26.349666538 +0000 UTC m=+66.352860379,LastTimestamp:2025-12-05 10:48:26.349666538 +0000 UTC m=+66.352860379,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:51:50.860959 master-0 kubenswrapper[24928]: I1205 10:51:50.860698 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:51:50.860959 master-0 kubenswrapper[24928]: I1205 10:51:50.860802 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:51:53.244761 master-0 kubenswrapper[24928]: I1205 10:51:53.244692 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-master-0" Dec 05 10:51:54.203776 master-0 kubenswrapper[24928]: I1205 10:51:54.203703 24928 scope.go:117] "RemoveContainer" containerID="d31f42cb2df60e0cf78918872183788f5baa6dfb0312a552f84381b8462fd7e0" Dec 05 10:51:54.905783 master-0 kubenswrapper[24928]: I1205 10:51:54.905538 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:51:54.905783 master-0 kubenswrapper[24928]: I1205 10:51:54.905772 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:51:55.171289 master-0 kubenswrapper[24928]: I1205 10:51:55.171114 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/3.log" Dec 05 10:51:55.171289 master-0 kubenswrapper[24928]: I1205 10:51:55.171215 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-lgn6v" event={"ID":"e27c0798-ec1c-43cd-b81b-f77f2f11ad0f","Type":"ContainerStarted","Data":"1350b871b2f533b59bac1b343c66f17d5af18b159632fce3313bed2ff7728e54"} Dec 05 10:51:59.936155 master-0 kubenswrapper[24928]: E1205 10:51:59.936007 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Timeout: request did not complete within requested timeout - context deadline exceeded" pod="openshift-etcd/etcd-master-0" Dec 05 10:52:00.204241 master-0 kubenswrapper[24928]: I1205 10:52:00.204083 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:52:00.204241 master-0 kubenswrapper[24928]: I1205 10:52:00.204122 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:52:00.860868 master-0 kubenswrapper[24928]: I1205 10:52:00.860799 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:52:00.861103 master-0 kubenswrapper[24928]: I1205 10:52:00.860871 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:52:04.905084 master-0 kubenswrapper[24928]: I1205 10:52:04.904977 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:52:04.905802 master-0 kubenswrapper[24928]: I1205 10:52:04.905097 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:52:10.860758 master-0 kubenswrapper[24928]: I1205 10:52:10.860678 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:52:10.860758 master-0 kubenswrapper[24928]: I1205 10:52:10.860738 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:52:14.904119 master-0 kubenswrapper[24928]: I1205 10:52:14.904010 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:52:14.904119 master-0 kubenswrapper[24928]: I1205 10:52:14.904090 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:52:20.166679 master-0 kubenswrapper[24928]: I1205 10:52:20.166582 24928 kubelet.go:1505] "Image garbage collection succeeded" Dec 05 10:52:20.860679 master-0 kubenswrapper[24928]: I1205 10:52:20.860606 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:52:20.860679 master-0 kubenswrapper[24928]: I1205 10:52:20.860676 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:52:23.078956 master-0 kubenswrapper[24928]: E1205 10:52:23.078769 24928 event.go:359] "Server rejected event (will not retry!)" err="Timeout: request did not complete within requested timeout - context deadline exceeded" event=< Dec 05 10:52:23.078956 master-0 kubenswrapper[24928]: &Event{ObjectMeta:{downloads-69cd4c69bf-d9jtn.187e4c0956ab1a2f openshift-console 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-console,Name:downloads-69cd4c69bf-d9jtn,UID:3c708092-6503-4a61-9230-617f8a1e2d19,APIVersion:v1,ResourceVersion:13063,FieldPath:spec.containers{download-server},},Reason:ProbeError,Message:Readiness probe error: Get "http://10.128.0.85:8080/": dial tcp 10.128.0.85:8080: connect: connection refused Dec 05 10:52:23.078956 master-0 kubenswrapper[24928]: body: Dec 05 10:52:23.078956 master-0 kubenswrapper[24928]: ,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:48:26.968283695 +0000 UTC m=+66.971477566,LastTimestamp:2025-12-05 10:48:26.968283695 +0000 UTC m=+66.971477566,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,} Dec 05 10:52:23.078956 master-0 kubenswrapper[24928]: > Dec 05 10:52:24.904563 master-0 kubenswrapper[24928]: I1205 10:52:24.904509 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:52:24.905321 master-0 kubenswrapper[24928]: I1205 10:52:24.905284 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:52:26.401515 master-0 kubenswrapper[24928]: I1205 10:52:26.401369 24928 status_manager.go:851] "Failed to get status for pod" podUID="ee0f69ec-1862-4016-bf9d-7edf6554f473" pod="openshift-etcd/installer-2-retry-1-master-0" err="the server was unable to return a response in the time allotted, but may still be processing the request (get pods installer-2-retry-1-master-0)" Dec 05 10:52:30.860929 master-0 kubenswrapper[24928]: I1205 10:52:30.860792 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:52:30.860929 master-0 kubenswrapper[24928]: I1205 10:52:30.860877 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:52:34.207515 master-0 kubenswrapper[24928]: E1205 10:52:34.207394 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Timeout: request did not complete within requested timeout - context deadline exceeded" pod="openshift-etcd/etcd-master-0" Dec 05 10:52:34.904594 master-0 kubenswrapper[24928]: I1205 10:52:34.904519 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:52:34.904848 master-0 kubenswrapper[24928]: I1205 10:52:34.904595 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:52:36.461536 master-0 kubenswrapper[24928]: I1205 10:52:36.461270 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-baremetal-operator-78f758c7b9-6t2gm_48bd1d86-a6f2-439f-ab04-6a9a442bec42/cluster-baremetal-operator/1.log" Dec 05 10:52:36.462513 master-0 kubenswrapper[24928]: I1205 10:52:36.462476 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-baremetal-operator-78f758c7b9-6t2gm_48bd1d86-a6f2-439f-ab04-6a9a442bec42/cluster-baremetal-operator/0.log" Dec 05 10:52:36.462661 master-0 kubenswrapper[24928]: I1205 10:52:36.462555 24928 generic.go:334] "Generic (PLEG): container finished" podID="48bd1d86-a6f2-439f-ab04-6a9a442bec42" containerID="10fa0576349384ec32341165d84160e0c9b5b5d6f5044617a559a0d068966dc5" exitCode=1 Dec 05 10:52:36.462661 master-0 kubenswrapper[24928]: I1205 10:52:36.462593 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" event={"ID":"48bd1d86-a6f2-439f-ab04-6a9a442bec42","Type":"ContainerDied","Data":"10fa0576349384ec32341165d84160e0c9b5b5d6f5044617a559a0d068966dc5"} Dec 05 10:52:36.462661 master-0 kubenswrapper[24928]: I1205 10:52:36.462639 24928 scope.go:117] "RemoveContainer" containerID="3a9cca203c845359b6a28fb821277d5c1db034d15078e96e3ff87b488109d4de" Dec 05 10:52:36.463912 master-0 kubenswrapper[24928]: I1205 10:52:36.463832 24928 scope.go:117] "RemoveContainer" containerID="10fa0576349384ec32341165d84160e0c9b5b5d6f5044617a559a0d068966dc5" Dec 05 10:52:36.464871 master-0 kubenswrapper[24928]: E1205 10:52:36.464714 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cluster-baremetal-operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cluster-baremetal-operator pod=cluster-baremetal-operator-78f758c7b9-6t2gm_openshift-machine-api(48bd1d86-a6f2-439f-ab04-6a9a442bec42)\"" pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" podUID="48bd1d86-a6f2-439f-ab04-6a9a442bec42" Dec 05 10:52:37.473293 master-0 kubenswrapper[24928]: I1205 10:52:37.473226 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-baremetal-operator-78f758c7b9-6t2gm_48bd1d86-a6f2-439f-ab04-6a9a442bec42/cluster-baremetal-operator/1.log" Dec 05 10:52:38.482494 master-0 kubenswrapper[24928]: I1205 10:52:38.482393 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-86f4478dbf-jqlt9_e0cbad64-72b9-4ad3-9a42-4183e93c9ba0/controller-manager/2.log" Dec 05 10:52:38.483394 master-0 kubenswrapper[24928]: I1205 10:52:38.482988 24928 generic.go:334] "Generic (PLEG): container finished" podID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerID="66ed166decce5e8377b96bf40238cb4562c34424e0d591957563972e8e5b494f" exitCode=255 Dec 05 10:52:38.483394 master-0 kubenswrapper[24928]: I1205 10:52:38.483034 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" event={"ID":"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0","Type":"ContainerDied","Data":"66ed166decce5e8377b96bf40238cb4562c34424e0d591957563972e8e5b494f"} Dec 05 10:52:38.483394 master-0 kubenswrapper[24928]: I1205 10:52:38.483076 24928 scope.go:117] "RemoveContainer" containerID="c3968a3a8ccd5212aba3d71d18db810366b8e5989f94f317e2256250850f0562" Dec 05 10:52:38.483746 master-0 kubenswrapper[24928]: I1205 10:52:38.483620 24928 scope.go:117] "RemoveContainer" containerID="66ed166decce5e8377b96bf40238cb4562c34424e0d591957563972e8e5b494f" Dec 05 10:52:38.483860 master-0 kubenswrapper[24928]: E1205 10:52:38.483825 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"controller-manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=controller-manager pod=controller-manager-86f4478dbf-jqlt9_openshift-controller-manager(e0cbad64-72b9-4ad3-9a42-4183e93c9ba0)\"" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" Dec 05 10:52:39.503719 master-0 kubenswrapper[24928]: I1205 10:52:39.503640 24928 generic.go:334] "Generic (PLEG): container finished" podID="88cccb5b-1ad1-4fab-b34c-90252794ee20" containerID="dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5" exitCode=0 Dec 05 10:52:39.504447 master-0 kubenswrapper[24928]: I1205 10:52:39.503726 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" event={"ID":"88cccb5b-1ad1-4fab-b34c-90252794ee20","Type":"ContainerDied","Data":"dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5"} Dec 05 10:52:39.505520 master-0 kubenswrapper[24928]: I1205 10:52:39.505495 24928 scope.go:117] "RemoveContainer" containerID="dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5" Dec 05 10:52:39.505811 master-0 kubenswrapper[24928]: I1205 10:52:39.505776 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-86f4478dbf-jqlt9_e0cbad64-72b9-4ad3-9a42-4183e93c9ba0/controller-manager/2.log" Dec 05 10:52:39.666638 master-0 kubenswrapper[24928]: I1205 10:52:39.666566 24928 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:52:39.667203 master-0 kubenswrapper[24928]: I1205 10:52:39.667171 24928 scope.go:117] "RemoveContainer" containerID="66ed166decce5e8377b96bf40238cb4562c34424e0d591957563972e8e5b494f" Dec 05 10:52:39.667397 master-0 kubenswrapper[24928]: E1205 10:52:39.667369 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"controller-manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=controller-manager pod=controller-manager-86f4478dbf-jqlt9_openshift-controller-manager(e0cbad64-72b9-4ad3-9a42-4183e93c9ba0)\"" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" Dec 05 10:52:40.515668 master-0 kubenswrapper[24928]: I1205 10:52:40.515604 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" event={"ID":"88cccb5b-1ad1-4fab-b34c-90252794ee20","Type":"ContainerStarted","Data":"35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0"} Dec 05 10:52:40.516246 master-0 kubenswrapper[24928]: I1205 10:52:40.515922 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:52:40.520001 master-0 kubenswrapper[24928]: I1205 10:52:40.519952 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:52:40.542903 master-0 kubenswrapper[24928]: I1205 10:52:40.542832 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:52:40.543344 master-0 kubenswrapper[24928]: I1205 10:52:40.543304 24928 scope.go:117] "RemoveContainer" containerID="66ed166decce5e8377b96bf40238cb4562c34424e0d591957563972e8e5b494f" Dec 05 10:52:40.543668 master-0 kubenswrapper[24928]: E1205 10:52:40.543630 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"controller-manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=controller-manager pod=controller-manager-86f4478dbf-jqlt9_openshift-controller-manager(e0cbad64-72b9-4ad3-9a42-4183e93c9ba0)\"" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" Dec 05 10:52:40.860555 master-0 kubenswrapper[24928]: I1205 10:52:40.860499 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:52:40.860555 master-0 kubenswrapper[24928]: I1205 10:52:40.860546 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:52:44.904105 master-0 kubenswrapper[24928]: I1205 10:52:44.904060 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:52:44.904889 master-0 kubenswrapper[24928]: I1205 10:52:44.904861 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:52:49.205312 master-0 kubenswrapper[24928]: I1205 10:52:49.205187 24928 scope.go:117] "RemoveContainer" containerID="10fa0576349384ec32341165d84160e0c9b5b5d6f5044617a559a0d068966dc5" Dec 05 10:52:50.584167 master-0 kubenswrapper[24928]: I1205 10:52:50.584100 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-baremetal-operator-78f758c7b9-6t2gm_48bd1d86-a6f2-439f-ab04-6a9a442bec42/cluster-baremetal-operator/1.log" Dec 05 10:52:50.584923 master-0 kubenswrapper[24928]: I1205 10:52:50.584761 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/cluster-baremetal-operator-78f758c7b9-6t2gm" event={"ID":"48bd1d86-a6f2-439f-ab04-6a9a442bec42","Type":"ContainerStarted","Data":"ae1501eed388badf70e9085359b931a089337ef1acf2f2c1b2f8991183fd343f"} Dec 05 10:52:50.860666 master-0 kubenswrapper[24928]: I1205 10:52:50.860455 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:52:50.860666 master-0 kubenswrapper[24928]: I1205 10:52:50.860565 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:52:52.600453 master-0 kubenswrapper[24928]: I1205 10:52:52.600304 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/cluster-policy-controller/1.log" Dec 05 10:52:52.601509 master-0 kubenswrapper[24928]: I1205 10:52:52.601490 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/1.log" Dec 05 10:52:52.603193 master-0 kubenswrapper[24928]: I1205 10:52:52.603167 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/0.log" Dec 05 10:52:52.603276 master-0 kubenswrapper[24928]: I1205 10:52:52.603223 24928 generic.go:334] "Generic (PLEG): container finished" podID="5219435a07a0220d41da97c4fb70abb1" containerID="cf784a713d53bc0541dfbf03c915a4f6b70148e2c7b42b695d4ce6650a26f08f" exitCode=255 Dec 05 10:52:52.603328 master-0 kubenswrapper[24928]: I1205 10:52:52.603281 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerDied","Data":"cf784a713d53bc0541dfbf03c915a4f6b70148e2c7b42b695d4ce6650a26f08f"} Dec 05 10:52:52.603405 master-0 kubenswrapper[24928]: I1205 10:52:52.603376 24928 scope.go:117] "RemoveContainer" containerID="a0c4bf77b56c9bf53793c3092e77173ade30a59f4de69720ced91f80c7a365c6" Dec 05 10:52:52.604094 master-0 kubenswrapper[24928]: I1205 10:52:52.604050 24928 scope.go:117] "RemoveContainer" containerID="cf784a713d53bc0541dfbf03c915a4f6b70148e2c7b42b695d4ce6650a26f08f" Dec 05 10:52:52.604374 master-0 kubenswrapper[24928]: E1205 10:52:52.604345 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-controller-manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-controller-manager pod=kube-controller-manager-master-0_openshift-kube-controller-manager(5219435a07a0220d41da97c4fb70abb1)\"" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" Dec 05 10:52:53.204673 master-0 kubenswrapper[24928]: I1205 10:52:53.204591 24928 scope.go:117] "RemoveContainer" containerID="66ed166decce5e8377b96bf40238cb4562c34424e0d591957563972e8e5b494f" Dec 05 10:52:53.298172 master-0 kubenswrapper[24928]: I1205 10:52:53.298091 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:52:53.613795 master-0 kubenswrapper[24928]: I1205 10:52:53.613718 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/cluster-policy-controller/1.log" Dec 05 10:52:53.615023 master-0 kubenswrapper[24928]: I1205 10:52:53.614974 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/1.log" Dec 05 10:52:53.616372 master-0 kubenswrapper[24928]: I1205 10:52:53.616313 24928 scope.go:117] "RemoveContainer" containerID="cf784a713d53bc0541dfbf03c915a4f6b70148e2c7b42b695d4ce6650a26f08f" Dec 05 10:52:53.616623 master-0 kubenswrapper[24928]: E1205 10:52:53.616595 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-controller-manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-controller-manager pod=kube-controller-manager-master-0_openshift-kube-controller-manager(5219435a07a0220d41da97c4fb70abb1)\"" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" Dec 05 10:52:53.617961 master-0 kubenswrapper[24928]: I1205 10:52:53.617914 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-86f4478dbf-jqlt9_e0cbad64-72b9-4ad3-9a42-4183e93c9ba0/controller-manager/2.log" Dec 05 10:52:53.618049 master-0 kubenswrapper[24928]: I1205 10:52:53.617997 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" event={"ID":"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0","Type":"ContainerStarted","Data":"019e6539df9821321d324eb135ee3e05f2d176e4391f04ca23a088fe40df360d"} Dec 05 10:52:53.618435 master-0 kubenswrapper[24928]: I1205 10:52:53.618374 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:52:53.623707 master-0 kubenswrapper[24928]: I1205 10:52:53.623610 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:52:54.904643 master-0 kubenswrapper[24928]: I1205 10:52:54.904533 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:52:54.905560 master-0 kubenswrapper[24928]: I1205 10:52:54.904643 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:52:57.081454 master-0 kubenswrapper[24928]: E1205 10:52:57.081255 24928 event.go:359] "Server rejected event (will not retry!)" err="Timeout: request did not complete within requested timeout - context deadline exceeded" event="&Event{ObjectMeta:{downloads-69cd4c69bf-d9jtn.187e4c0956abfdc4 openshift-console 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-console,Name:downloads-69cd4c69bf-d9jtn,UID:3c708092-6503-4a61-9230-617f8a1e2d19,APIVersion:v1,ResourceVersion:13063,FieldPath:spec.containers{download-server},},Reason:Unhealthy,Message:Readiness probe failed: Get \"http://10.128.0.85:8080/\": dial tcp 10.128.0.85:8080: connect: connection refused,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:48:26.968341956 +0000 UTC m=+66.971535817,LastTimestamp:2025-12-05 10:48:26.968341956 +0000 UTC m=+66.971535817,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:53:00.860709 master-0 kubenswrapper[24928]: I1205 10:53:00.860641 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:53:00.860709 master-0 kubenswrapper[24928]: I1205 10:53:00.860701 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:53:00.861453 master-0 kubenswrapper[24928]: I1205 10:53:00.860750 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:53:00.861453 master-0 kubenswrapper[24928]: I1205 10:53:00.861305 24928 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="console" containerStatusID={"Type":"cri-o","ID":"61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2"} pod="openshift-console/console-74f96dcf4d-9gskd" containerMessage="Container console failed startup probe, will be restarted" Dec 05 10:53:01.304315 master-0 kubenswrapper[24928]: I1205 10:53:01.304238 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:01.304894 master-0 kubenswrapper[24928]: I1205 10:53:01.304860 24928 scope.go:117] "RemoveContainer" containerID="cf784a713d53bc0541dfbf03c915a4f6b70148e2c7b42b695d4ce6650a26f08f" Dec 05 10:53:01.305159 master-0 kubenswrapper[24928]: E1205 10:53:01.305119 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-controller-manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-controller-manager pod=kube-controller-manager-master-0_openshift-kube-controller-manager(5219435a07a0220d41da97c4fb70abb1)\"" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" Dec 05 10:53:01.407871 master-0 kubenswrapper[24928]: I1205 10:53:01.407761 24928 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:01.683477 master-0 kubenswrapper[24928]: I1205 10:53:01.683244 24928 scope.go:117] "RemoveContainer" containerID="cf784a713d53bc0541dfbf03c915a4f6b70148e2c7b42b695d4ce6650a26f08f" Dec 05 10:53:01.683832 master-0 kubenswrapper[24928]: E1205 10:53:01.683720 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-controller-manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-controller-manager pod=kube-controller-manager-master-0_openshift-kube-controller-manager(5219435a07a0220d41da97c4fb70abb1)\"" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" Dec 05 10:53:04.904202 master-0 kubenswrapper[24928]: I1205 10:53:04.904129 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:53:04.905365 master-0 kubenswrapper[24928]: I1205 10:53:04.904214 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:53:04.905365 master-0 kubenswrapper[24928]: I1205 10:53:04.904287 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:53:04.906127 master-0 kubenswrapper[24928]: I1205 10:53:04.906040 24928 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="console" containerStatusID={"Type":"cri-o","ID":"4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02"} pod="openshift-console/console-79cdddb8b4-mwjwx" containerMessage="Container console failed startup probe, will be restarted" Dec 05 10:53:05.922503 master-0 kubenswrapper[24928]: E1205 10:53:05.922399 24928 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command 'sleep 25' exited with 137: " execCommand=["sleep","25"] containerName="console" pod="openshift-console/console-74f96dcf4d-9gskd" message="" Dec 05 10:53:05.922503 master-0 kubenswrapper[24928]: E1205 10:53:05.922499 24928 kuberuntime_container.go:691] "PreStop hook failed" err="command 'sleep 25' exited with 137: " pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" containerID="cri-o://61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2" Dec 05 10:53:05.922978 master-0 kubenswrapper[24928]: I1205 10:53:05.922556 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" containerID="cri-o://61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2" gracePeriod=35 Dec 05 10:53:06.339266 master-0 kubenswrapper[24928]: E1205 10:53:06.339206 24928 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err="command 'sleep 25' exited with 137: " execCommand=["sleep","25"] containerName="console" pod="openshift-console/console-79cdddb8b4-mwjwx" message="" Dec 05 10:53:06.339266 master-0 kubenswrapper[24928]: E1205 10:53:06.339260 24928 kuberuntime_container.go:691] "PreStop hook failed" err="command 'sleep 25' exited with 137: " pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" containerID="cri-o://4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02" Dec 05 10:53:06.339379 master-0 kubenswrapper[24928]: I1205 10:53:06.339311 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" containerID="cri-o://4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02" gracePeriod=39 Dec 05 10:53:06.734255 master-0 kubenswrapper[24928]: I1205 10:53:06.734051 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-79cdddb8b4-mwjwx_ac38876d-122b-4146-babf-722930c2c4ae/console/0.log" Dec 05 10:53:06.734255 master-0 kubenswrapper[24928]: I1205 10:53:06.734128 24928 generic.go:334] "Generic (PLEG): container finished" podID="ac38876d-122b-4146-babf-722930c2c4ae" containerID="4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02" exitCode=255 Dec 05 10:53:06.734877 master-0 kubenswrapper[24928]: I1205 10:53:06.734250 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79cdddb8b4-mwjwx" event={"ID":"ac38876d-122b-4146-babf-722930c2c4ae","Type":"ContainerDied","Data":"4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02"} Dec 05 10:53:06.734877 master-0 kubenswrapper[24928]: I1205 10:53:06.734324 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79cdddb8b4-mwjwx" event={"ID":"ac38876d-122b-4146-babf-722930c2c4ae","Type":"ContainerStarted","Data":"cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320"} Dec 05 10:53:06.737556 master-0 kubenswrapper[24928]: I1205 10:53:06.737471 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74f96dcf4d-9gskd_3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527/console/0.log" Dec 05 10:53:06.737556 master-0 kubenswrapper[24928]: I1205 10:53:06.737509 24928 generic.go:334] "Generic (PLEG): container finished" podID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerID="61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2" exitCode=255 Dec 05 10:53:06.737556 master-0 kubenswrapper[24928]: I1205 10:53:06.737546 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74f96dcf4d-9gskd" event={"ID":"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527","Type":"ContainerDied","Data":"61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2"} Dec 05 10:53:06.737862 master-0 kubenswrapper[24928]: I1205 10:53:06.737579 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74f96dcf4d-9gskd" event={"ID":"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527","Type":"ContainerStarted","Data":"ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311"} Dec 05 10:53:10.631175 master-0 kubenswrapper[24928]: I1205 10:53:10.631072 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/installer-4-master-0"] Dec 05 10:53:10.703521 master-0 kubenswrapper[24928]: I1205 10:53:10.695195 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-77b5b8969c-5clks"] Dec 05 10:53:10.703521 master-0 kubenswrapper[24928]: I1205 10:53:10.699947 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-77b5b8969c-5clks"] Dec 05 10:53:10.722104 master-0 kubenswrapper[24928]: I1205 10:53:10.721936 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/installer-4-master-0"] Dec 05 10:53:10.722562 master-0 kubenswrapper[24928]: E1205 10:53:10.722546 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75d79a3b-db06-4b9d-99ca-9b1a12dd44da" containerName="oauth-openshift" Dec 05 10:53:10.722645 master-0 kubenswrapper[24928]: I1205 10:53:10.722633 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="75d79a3b-db06-4b9d-99ca-9b1a12dd44da" containerName="oauth-openshift" Dec 05 10:53:10.722742 master-0 kubenswrapper[24928]: E1205 10:53:10.722731 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3" containerName="installer" Dec 05 10:53:10.722803 master-0 kubenswrapper[24928]: I1205 10:53:10.722793 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3" containerName="installer" Dec 05 10:53:10.722876 master-0 kubenswrapper[24928]: E1205 10:53:10.722866 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee0f69ec-1862-4016-bf9d-7edf6554f473" containerName="installer" Dec 05 10:53:10.722931 master-0 kubenswrapper[24928]: I1205 10:53:10.722922 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0f69ec-1862-4016-bf9d-7edf6554f473" containerName="installer" Dec 05 10:53:10.723116 master-0 kubenswrapper[24928]: I1205 10:53:10.723104 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee0f69ec-1862-4016-bf9d-7edf6554f473" containerName="installer" Dec 05 10:53:10.723187 master-0 kubenswrapper[24928]: I1205 10:53:10.723177 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="75d79a3b-db06-4b9d-99ca-9b1a12dd44da" containerName="oauth-openshift" Dec 05 10:53:10.723250 master-0 kubenswrapper[24928]: I1205 10:53:10.723241 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3" containerName="installer" Dec 05 10:53:10.723733 master-0 kubenswrapper[24928]: I1205 10:53:10.723717 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:10.728718 master-0 kubenswrapper[24928]: I1205 10:53:10.728681 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-pcdf6" Dec 05 10:53:10.728995 master-0 kubenswrapper[24928]: I1205 10:53:10.728777 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 05 10:53:10.751197 master-0 kubenswrapper[24928]: I1205 10:53:10.750926 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/installer-4-master-0"] Dec 05 10:53:10.774535 master-0 kubenswrapper[24928]: I1205 10:53:10.774463 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/installer-4-master-0"] Dec 05 10:53:10.781038 master-0 kubenswrapper[24928]: I1205 10:53:10.780956 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/installer-4-master-0"] Dec 05 10:53:10.842024 master-0 kubenswrapper[24928]: I1205 10:53:10.841688 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-var-lock\") pod \"installer-4-master-0\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:10.842024 master-0 kubenswrapper[24928]: I1205 10:53:10.841821 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52a48f6c-b05f-4b70-96cc-378e37928c94-kube-api-access\") pod \"installer-4-master-0\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:10.842024 master-0 kubenswrapper[24928]: I1205 10:53:10.841861 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-kubelet-dir\") pod \"installer-4-master-0\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:10.862526 master-0 kubenswrapper[24928]: I1205 10:53:10.862456 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:53:10.862814 master-0 kubenswrapper[24928]: I1205 10:53:10.862759 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:53:10.863006 master-0 kubenswrapper[24928]: I1205 10:53:10.862962 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:53:10.863073 master-0 kubenswrapper[24928]: I1205 10:53:10.862795 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:53:10.943059 master-0 kubenswrapper[24928]: I1205 10:53:10.942908 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-kubelet-dir\") pod \"installer-4-master-0\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:10.943291 master-0 kubenswrapper[24928]: I1205 10:53:10.943225 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-kubelet-dir\") pod \"installer-4-master-0\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:10.943501 master-0 kubenswrapper[24928]: I1205 10:53:10.943453 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-var-lock\") pod \"installer-4-master-0\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:10.943619 master-0 kubenswrapper[24928]: I1205 10:53:10.943572 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-var-lock\") pod \"installer-4-master-0\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:10.943667 master-0 kubenswrapper[24928]: I1205 10:53:10.943639 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52a48f6c-b05f-4b70-96cc-378e37928c94-kube-api-access\") pod \"installer-4-master-0\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:11.057115 master-0 kubenswrapper[24928]: I1205 10:53:11.056996 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-69cd4c69bf-d9jtn" podStartSLOduration=285.899578819 podStartE2EDuration="5m14.056981006s" podCreationTimestamp="2025-12-05 10:47:57 +0000 UTC" firstStartedPulling="2025-12-05 10:47:58.031331725 +0000 UTC m=+38.034525576" lastFinishedPulling="2025-12-05 10:48:26.188733912 +0000 UTC m=+66.191927763" observedRunningTime="2025-12-05 10:53:11.054465329 +0000 UTC m=+351.057659190" watchObservedRunningTime="2025-12-05 10:53:11.056981006 +0000 UTC m=+351.060174847" Dec 05 10:53:12.212823 master-0 kubenswrapper[24928]: I1205 10:53:12.212749 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75d79a3b-db06-4b9d-99ca-9b1a12dd44da" path="/var/lib/kubelet/pods/75d79a3b-db06-4b9d-99ca-9b1a12dd44da/volumes" Dec 05 10:53:12.213627 master-0 kubenswrapper[24928]: I1205 10:53:12.213594 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3" path="/var/lib/kubelet/pods/f7d7bcf4-17f9-49c5-bcc0-10ca50ee8ba3/volumes" Dec 05 10:53:12.599108 master-0 kubenswrapper[24928]: I1205 10:53:12.599054 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-5-master-0"] Dec 05 10:53:12.599954 master-0 kubenswrapper[24928]: I1205 10:53:12.599930 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:12.602049 master-0 kubenswrapper[24928]: I1205 10:53:12.601987 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 10:53:12.602398 master-0 kubenswrapper[24928]: I1205 10:53:12.602356 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-grfdw" Dec 05 10:53:12.614051 master-0 kubenswrapper[24928]: I1205 10:53:12.614010 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-5-master-0"] Dec 05 10:53:12.668271 master-0 kubenswrapper[24928]: I1205 10:53:12.668194 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kubelet-dir\") pod \"installer-5-master-0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:12.668500 master-0 kubenswrapper[24928]: I1205 10:53:12.668410 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-var-lock\") pod \"installer-5-master-0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:12.668718 master-0 kubenswrapper[24928]: I1205 10:53:12.668678 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kube-api-access\") pod \"installer-5-master-0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:12.770741 master-0 kubenswrapper[24928]: I1205 10:53:12.770640 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kube-api-access\") pod \"installer-5-master-0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:12.771179 master-0 kubenswrapper[24928]: I1205 10:53:12.771139 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kubelet-dir\") pod \"installer-5-master-0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:12.771276 master-0 kubenswrapper[24928]: I1205 10:53:12.771197 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-var-lock\") pod \"installer-5-master-0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:12.771408 master-0 kubenswrapper[24928]: I1205 10:53:12.771360 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kubelet-dir\") pod \"installer-5-master-0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:12.771533 master-0 kubenswrapper[24928]: I1205 10:53:12.771477 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-var-lock\") pod \"installer-5-master-0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:13.662164 master-0 kubenswrapper[24928]: I1205 10:53:13.659672 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52a48f6c-b05f-4b70-96cc-378e37928c94-kube-api-access\") pod \"installer-4-master-0\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:13.662164 master-0 kubenswrapper[24928]: I1205 10:53:13.661137 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kube-api-access\") pod \"installer-5-master-0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:13.756625 master-0 kubenswrapper[24928]: I1205 10:53:13.756555 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:13.822450 master-0 kubenswrapper[24928]: I1205 10:53:13.821765 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:14.335633 master-0 kubenswrapper[24928]: I1205 10:53:14.335572 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/installer-4-master-0"] Dec 05 10:53:14.340752 master-0 kubenswrapper[24928]: I1205 10:53:14.340633 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-5-master-0"] Dec 05 10:53:14.344150 master-0 kubenswrapper[24928]: W1205 10:53:14.344096 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod52a48f6c_b05f_4b70_96cc_378e37928c94.slice/crio-83f85cf1dc2a5ddb90cf7d96dd9dbdd7fb07b05609bc0168ff762882c11c6a8c WatchSource:0}: Error finding container 83f85cf1dc2a5ddb90cf7d96dd9dbdd7fb07b05609bc0168ff762882c11c6a8c: Status 404 returned error can't find the container with id 83f85cf1dc2a5ddb90cf7d96dd9dbdd7fb07b05609bc0168ff762882c11c6a8c Dec 05 10:53:14.814287 master-0 kubenswrapper[24928]: I1205 10:53:14.814225 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-4-master-0" event={"ID":"52a48f6c-b05f-4b70-96cc-378e37928c94","Type":"ContainerStarted","Data":"d539da331de2ee7e892ad095acecd0bf1191a2be47d514f0e5423eec137d52fd"} Dec 05 10:53:14.814287 master-0 kubenswrapper[24928]: I1205 10:53:14.814283 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-4-master-0" event={"ID":"52a48f6c-b05f-4b70-96cc-378e37928c94","Type":"ContainerStarted","Data":"83f85cf1dc2a5ddb90cf7d96dd9dbdd7fb07b05609bc0168ff762882c11c6a8c"} Dec 05 10:53:14.816249 master-0 kubenswrapper[24928]: I1205 10:53:14.816182 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-5-master-0" event={"ID":"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0","Type":"ContainerStarted","Data":"24059d1d31317d09c58db0a6c95ce43a0078bec9901eb497277dd670e3b3751e"} Dec 05 10:53:14.834807 master-0 kubenswrapper[24928]: I1205 10:53:14.834703 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-5-master-0" podStartSLOduration=2.834678624 podStartE2EDuration="2.834678624s" podCreationTimestamp="2025-12-05 10:53:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:53:14.832769059 +0000 UTC m=+354.835962910" watchObservedRunningTime="2025-12-05 10:53:14.834678624 +0000 UTC m=+354.837872475" Dec 05 10:53:14.903992 master-0 kubenswrapper[24928]: I1205 10:53:14.903846 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:53:14.903992 master-0 kubenswrapper[24928]: I1205 10:53:14.903930 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:53:14.904428 master-0 kubenswrapper[24928]: I1205 10:53:14.904370 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:53:14.904505 master-0 kubenswrapper[24928]: I1205 10:53:14.904460 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:53:15.824743 master-0 kubenswrapper[24928]: I1205 10:53:15.824680 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-5-master-0" event={"ID":"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0","Type":"ContainerStarted","Data":"8e36c34ac1ad0faf4333c95bccd24ab7f0d2a8d130aee4d157dbd04cb3cc5423"} Dec 05 10:53:15.851016 master-0 kubenswrapper[24928]: I1205 10:53:15.849941 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/installer-4-master-0" podStartSLOduration=5.849925232 podStartE2EDuration="5.849925232s" podCreationTimestamp="2025-12-05 10:53:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:53:15.848250983 +0000 UTC m=+355.851444834" watchObservedRunningTime="2025-12-05 10:53:15.849925232 +0000 UTC m=+355.853119083" Dec 05 10:53:16.204667 master-0 kubenswrapper[24928]: I1205 10:53:16.204537 24928 scope.go:117] "RemoveContainer" containerID="cf784a713d53bc0541dfbf03c915a4f6b70148e2c7b42b695d4ce6650a26f08f" Dec 05 10:53:16.833134 master-0 kubenswrapper[24928]: I1205 10:53:16.833087 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/cluster-policy-controller/1.log" Dec 05 10:53:16.834057 master-0 kubenswrapper[24928]: I1205 10:53:16.834031 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/1.log" Dec 05 10:53:16.835011 master-0 kubenswrapper[24928]: I1205 10:53:16.834960 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"5219435a07a0220d41da97c4fb70abb1","Type":"ContainerStarted","Data":"5e1cd037da3a72a86565cde7e25ecd2da3f7f8cf8ea4a83328e0ab5b87b966ad"} Dec 05 10:53:20.861796 master-0 kubenswrapper[24928]: I1205 10:53:20.861747 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:53:20.862373 master-0 kubenswrapper[24928]: I1205 10:53:20.861799 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:53:21.304328 master-0 kubenswrapper[24928]: I1205 10:53:21.304279 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:21.307473 master-0 kubenswrapper[24928]: I1205 10:53:21.307413 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:21.863656 master-0 kubenswrapper[24928]: I1205 10:53:21.863564 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:24.904194 master-0 kubenswrapper[24928]: I1205 10:53:24.904114 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:53:24.904194 master-0 kubenswrapper[24928]: I1205 10:53:24.904178 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:53:30.861076 master-0 kubenswrapper[24928]: I1205 10:53:30.861004 24928 patch_prober.go:28] interesting pod/console-74f96dcf4d-9gskd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" start-of-body= Dec 05 10:53:30.862284 master-0 kubenswrapper[24928]: I1205 10:53:30.861091 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" probeResult="failure" output="Get \"https://10.128.0.86:8443/health\": dial tcp 10.128.0.86:8443: connect: connection refused" Dec 05 10:53:33.302670 master-0 kubenswrapper[24928]: I1205 10:53:33.302587 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:34.149485 master-0 kubenswrapper[24928]: I1205 10:53:34.146095 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/metrics-server-7c46d76dff-z8d8z"] Dec 05 10:53:34.149485 master-0 kubenswrapper[24928]: I1205 10:53:34.146365 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" podUID="3aa9a063-322b-4dc6-a724-05a66026160b" containerName="metrics-server" containerID="cri-o://20fd7ecb75706506f53b22d6ef8c2f7cefab03481f4f534c87bae4cf23aa6438" gracePeriod=170 Dec 05 10:53:34.156111 master-0 kubenswrapper[24928]: I1205 10:53:34.152019 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-74f96dcf4d-9gskd"] Dec 05 10:53:34.156307 master-0 kubenswrapper[24928]: I1205 10:53:34.156197 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-86f4478dbf-jqlt9"] Dec 05 10:53:34.156576 master-0 kubenswrapper[24928]: I1205 10:53:34.156476 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" containerID="cri-o://019e6539df9821321d324eb135ee3e05f2d176e4391f04ca23a088fe40df360d" gracePeriod=30 Dec 05 10:53:34.269447 master-0 kubenswrapper[24928]: I1205 10:53:34.269137 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s"] Dec 05 10:53:34.269447 master-0 kubenswrapper[24928]: I1205 10:53:34.269377 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" podUID="88cccb5b-1ad1-4fab-b34c-90252794ee20" containerName="route-controller-manager" containerID="cri-o://35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0" gracePeriod=30 Dec 05 10:53:34.885870 master-0 kubenswrapper[24928]: I1205 10:53:34.885811 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:53:34.903677 master-0 kubenswrapper[24928]: I1205 10:53:34.903607 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:53:34.903886 master-0 kubenswrapper[24928]: I1205 10:53:34.903702 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:53:34.951307 master-0 kubenswrapper[24928]: I1205 10:53:34.951224 24928 generic.go:334] "Generic (PLEG): container finished" podID="88cccb5b-1ad1-4fab-b34c-90252794ee20" containerID="35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0" exitCode=0 Dec 05 10:53:34.951307 master-0 kubenswrapper[24928]: I1205 10:53:34.951275 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" event={"ID":"88cccb5b-1ad1-4fab-b34c-90252794ee20","Type":"ContainerDied","Data":"35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0"} Dec 05 10:53:34.951603 master-0 kubenswrapper[24928]: I1205 10:53:34.951302 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" Dec 05 10:53:34.951603 master-0 kubenswrapper[24928]: I1205 10:53:34.951508 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s" event={"ID":"88cccb5b-1ad1-4fab-b34c-90252794ee20","Type":"ContainerDied","Data":"1bde7edcbb5c913afcdfed1e9b165ffe883b909863e89e771b54560c18c0ac2a"} Dec 05 10:53:34.951603 master-0 kubenswrapper[24928]: I1205 10:53:34.951535 24928 scope.go:117] "RemoveContainer" containerID="35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0" Dec 05 10:53:34.955924 master-0 kubenswrapper[24928]: I1205 10:53:34.955885 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-86f4478dbf-jqlt9_e0cbad64-72b9-4ad3-9a42-4183e93c9ba0/controller-manager/2.log" Dec 05 10:53:34.956008 master-0 kubenswrapper[24928]: I1205 10:53:34.955947 24928 generic.go:334] "Generic (PLEG): container finished" podID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerID="019e6539df9821321d324eb135ee3e05f2d176e4391f04ca23a088fe40df360d" exitCode=0 Dec 05 10:53:34.956008 master-0 kubenswrapper[24928]: I1205 10:53:34.955976 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" event={"ID":"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0","Type":"ContainerDied","Data":"019e6539df9821321d324eb135ee3e05f2d176e4391f04ca23a088fe40df360d"} Dec 05 10:53:34.970083 master-0 kubenswrapper[24928]: I1205 10:53:34.969883 24928 scope.go:117] "RemoveContainer" containerID="dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5" Dec 05 10:53:34.993228 master-0 kubenswrapper[24928]: I1205 10:53:34.992019 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-client-ca\") pod \"88cccb5b-1ad1-4fab-b34c-90252794ee20\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " Dec 05 10:53:34.993228 master-0 kubenswrapper[24928]: I1205 10:53:34.992135 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88cccb5b-1ad1-4fab-b34c-90252794ee20-serving-cert\") pod \"88cccb5b-1ad1-4fab-b34c-90252794ee20\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " Dec 05 10:53:34.993228 master-0 kubenswrapper[24928]: I1205 10:53:34.992206 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-config\") pod \"88cccb5b-1ad1-4fab-b34c-90252794ee20\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " Dec 05 10:53:34.993228 master-0 kubenswrapper[24928]: I1205 10:53:34.992269 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4lbg\" (UniqueName: \"kubernetes.io/projected/88cccb5b-1ad1-4fab-b34c-90252794ee20-kube-api-access-f4lbg\") pod \"88cccb5b-1ad1-4fab-b34c-90252794ee20\" (UID: \"88cccb5b-1ad1-4fab-b34c-90252794ee20\") " Dec 05 10:53:34.993688 master-0 kubenswrapper[24928]: I1205 10:53:34.993530 24928 scope.go:117] "RemoveContainer" containerID="35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0" Dec 05 10:53:34.994241 master-0 kubenswrapper[24928]: I1205 10:53:34.994197 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-config" (OuterVolumeSpecName: "config") pod "88cccb5b-1ad1-4fab-b34c-90252794ee20" (UID: "88cccb5b-1ad1-4fab-b34c-90252794ee20"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:53:34.994318 master-0 kubenswrapper[24928]: E1205 10:53:34.994287 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0\": container with ID starting with 35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0 not found: ID does not exist" containerID="35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0" Dec 05 10:53:34.994372 master-0 kubenswrapper[24928]: I1205 10:53:34.994309 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0"} err="failed to get container status \"35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0\": rpc error: code = NotFound desc = could not find container \"35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0\": container with ID starting with 35fb8542a2180e3d0751ace9eb78961cb0fcf16182365039ada70ac46cb8dce0 not found: ID does not exist" Dec 05 10:53:34.994372 master-0 kubenswrapper[24928]: I1205 10:53:34.994329 24928 scope.go:117] "RemoveContainer" containerID="dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5" Dec 05 10:53:34.994975 master-0 kubenswrapper[24928]: E1205 10:53:34.994928 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5\": container with ID starting with dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5 not found: ID does not exist" containerID="dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5" Dec 05 10:53:34.995073 master-0 kubenswrapper[24928]: I1205 10:53:34.994985 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5"} err="failed to get container status \"dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5\": rpc error: code = NotFound desc = could not find container \"dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5\": container with ID starting with dc84fa8bc44e802f284ac21a42e77ff03fe8831109e5faaadb6973507d39cdc5 not found: ID does not exist" Dec 05 10:53:34.995073 master-0 kubenswrapper[24928]: I1205 10:53:34.995017 24928 scope.go:117] "RemoveContainer" containerID="66ed166decce5e8377b96bf40238cb4562c34424e0d591957563972e8e5b494f" Dec 05 10:53:34.995494 master-0 kubenswrapper[24928]: I1205 10:53:34.995292 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-client-ca" (OuterVolumeSpecName: "client-ca") pod "88cccb5b-1ad1-4fab-b34c-90252794ee20" (UID: "88cccb5b-1ad1-4fab-b34c-90252794ee20"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:53:34.996124 master-0 kubenswrapper[24928]: I1205 10:53:34.996076 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88cccb5b-1ad1-4fab-b34c-90252794ee20-kube-api-access-f4lbg" (OuterVolumeSpecName: "kube-api-access-f4lbg") pod "88cccb5b-1ad1-4fab-b34c-90252794ee20" (UID: "88cccb5b-1ad1-4fab-b34c-90252794ee20"). InnerVolumeSpecName "kube-api-access-f4lbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:53:34.998962 master-0 kubenswrapper[24928]: I1205 10:53:34.998841 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88cccb5b-1ad1-4fab-b34c-90252794ee20-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "88cccb5b-1ad1-4fab-b34c-90252794ee20" (UID: "88cccb5b-1ad1-4fab-b34c-90252794ee20"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:53:35.094509 master-0 kubenswrapper[24928]: I1205 10:53:35.094448 24928 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-client-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:35.094509 master-0 kubenswrapper[24928]: I1205 10:53:35.094497 24928 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88cccb5b-1ad1-4fab-b34c-90252794ee20-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:35.094509 master-0 kubenswrapper[24928]: I1205 10:53:35.094511 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88cccb5b-1ad1-4fab-b34c-90252794ee20-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:35.094702 master-0 kubenswrapper[24928]: I1205 10:53:35.094526 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4lbg\" (UniqueName: \"kubernetes.io/projected/88cccb5b-1ad1-4fab-b34c-90252794ee20-kube-api-access-f4lbg\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:35.100846 master-0 kubenswrapper[24928]: I1205 10:53:35.100812 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:53:35.195374 master-0 kubenswrapper[24928]: I1205 10:53:35.195312 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-serving-cert\") pod \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " Dec 05 10:53:35.195374 master-0 kubenswrapper[24928]: I1205 10:53:35.195384 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-client-ca\") pod \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " Dec 05 10:53:35.195853 master-0 kubenswrapper[24928]: I1205 10:53:35.195460 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-config\") pod \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " Dec 05 10:53:35.195853 master-0 kubenswrapper[24928]: I1205 10:53:35.195532 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95jmp\" (UniqueName: \"kubernetes.io/projected/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-kube-api-access-95jmp\") pod \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " Dec 05 10:53:35.195853 master-0 kubenswrapper[24928]: I1205 10:53:35.195562 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-proxy-ca-bundles\") pod \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\" (UID: \"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0\") " Dec 05 10:53:35.196493 master-0 kubenswrapper[24928]: I1205 10:53:35.196452 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-client-ca" (OuterVolumeSpecName: "client-ca") pod "e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" (UID: "e0cbad64-72b9-4ad3-9a42-4183e93c9ba0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:53:35.196893 master-0 kubenswrapper[24928]: I1205 10:53:35.196661 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-config" (OuterVolumeSpecName: "config") pod "e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" (UID: "e0cbad64-72b9-4ad3-9a42-4183e93c9ba0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:53:35.196893 master-0 kubenswrapper[24928]: I1205 10:53:35.196808 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" (UID: "e0cbad64-72b9-4ad3-9a42-4183e93c9ba0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:53:35.198366 master-0 kubenswrapper[24928]: I1205 10:53:35.198336 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" (UID: "e0cbad64-72b9-4ad3-9a42-4183e93c9ba0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:53:35.198723 master-0 kubenswrapper[24928]: I1205 10:53:35.198695 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-kube-api-access-95jmp" (OuterVolumeSpecName: "kube-api-access-95jmp") pod "e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" (UID: "e0cbad64-72b9-4ad3-9a42-4183e93c9ba0"). InnerVolumeSpecName "kube-api-access-95jmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:53:35.297891 master-0 kubenswrapper[24928]: I1205 10:53:35.297853 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95jmp\" (UniqueName: \"kubernetes.io/projected/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-kube-api-access-95jmp\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:35.298022 master-0 kubenswrapper[24928]: I1205 10:53:35.298008 24928 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-proxy-ca-bundles\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:35.298118 master-0 kubenswrapper[24928]: I1205 10:53:35.298104 24928 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:35.298203 master-0 kubenswrapper[24928]: I1205 10:53:35.298190 24928 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-client-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:35.298269 master-0 kubenswrapper[24928]: I1205 10:53:35.298259 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:35.321071 master-0 kubenswrapper[24928]: I1205 10:53:35.320998 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s"] Dec 05 10:53:35.324932 master-0 kubenswrapper[24928]: I1205 10:53:35.324873 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c7946c9c4-hq97s"] Dec 05 10:53:35.964350 master-0 kubenswrapper[24928]: I1205 10:53:35.964277 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" event={"ID":"e0cbad64-72b9-4ad3-9a42-4183e93c9ba0","Type":"ContainerDied","Data":"0f19c2361e4856d55a1f7cfcf2bbf2e6aae6be0b4b861e0e68ba7772ccbc2487"} Dec 05 10:53:35.964350 master-0 kubenswrapper[24928]: I1205 10:53:35.964346 24928 scope.go:117] "RemoveContainer" containerID="019e6539df9821321d324eb135ee3e05f2d176e4391f04ca23a088fe40df360d" Dec 05 10:53:35.965024 master-0 kubenswrapper[24928]: I1205 10:53:35.964413 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-86f4478dbf-jqlt9" Dec 05 10:53:35.998598 master-0 kubenswrapper[24928]: I1205 10:53:35.998527 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-86f4478dbf-jqlt9"] Dec 05 10:53:36.003339 master-0 kubenswrapper[24928]: I1205 10:53:36.003275 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-86f4478dbf-jqlt9"] Dec 05 10:53:36.212032 master-0 kubenswrapper[24928]: I1205 10:53:36.211978 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88cccb5b-1ad1-4fab-b34c-90252794ee20" path="/var/lib/kubelet/pods/88cccb5b-1ad1-4fab-b34c-90252794ee20/volumes" Dec 05 10:53:36.212695 master-0 kubenswrapper[24928]: I1205 10:53:36.212670 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" path="/var/lib/kubelet/pods/e0cbad64-72b9-4ad3-9a42-4183e93c9ba0/volumes" Dec 05 10:53:44.904059 master-0 kubenswrapper[24928]: I1205 10:53:44.903979 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:53:44.904641 master-0 kubenswrapper[24928]: I1205 10:53:44.904057 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:53:47.393008 master-0 kubenswrapper[24928]: I1205 10:53:47.392944 24928 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-master-0"] Dec 05 10:53:47.393629 master-0 kubenswrapper[24928]: I1205 10:53:47.393224 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager-cert-syncer" containerID="cri-o://47bef9607325414e871a594b06d4250139a33f93aa8b2368ce9838e31d093051" gracePeriod=30 Dec 05 10:53:47.393629 master-0 kubenswrapper[24928]: I1205 10:53:47.393297 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" containerID="cri-o://b27b1cff74cf1fa7645c9d670ba8742571627f768b820d46bb086bfbeb3ea29b" gracePeriod=30 Dec 05 10:53:47.393629 master-0 kubenswrapper[24928]: I1205 10:53:47.393313 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager-recovery-controller" containerID="cri-o://50e85593f8bade43c99e607efe092eba239c2c489a561c3d4ee019a9a7b227cb" gracePeriod=30 Dec 05 10:53:47.393629 master-0 kubenswrapper[24928]: I1205 10:53:47.393412 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager" containerID="cri-o://5e1cd037da3a72a86565cde7e25ecd2da3f7f8cf8ea4a83328e0ab5b87b966ad" gracePeriod=30 Dec 05 10:53:47.394693 master-0 kubenswrapper[24928]: I1205 10:53:47.394602 24928 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-master-0"] Dec 05 10:53:47.394979 master-0 kubenswrapper[24928]: E1205 10:53:47.394941 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager-recovery-controller" Dec 05 10:53:47.394979 master-0 kubenswrapper[24928]: I1205 10:53:47.394970 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager-recovery-controller" Dec 05 10:53:47.395053 master-0 kubenswrapper[24928]: E1205 10:53:47.395000 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager-cert-syncer" Dec 05 10:53:47.395053 master-0 kubenswrapper[24928]: I1205 10:53:47.395009 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager-cert-syncer" Dec 05 10:53:47.395053 master-0 kubenswrapper[24928]: E1205 10:53:47.395035 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.395053 master-0 kubenswrapper[24928]: I1205 10:53:47.395044 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.395165 master-0 kubenswrapper[24928]: E1205 10:53:47.395061 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.395165 master-0 kubenswrapper[24928]: I1205 10:53:47.395071 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.395165 master-0 kubenswrapper[24928]: E1205 10:53:47.395083 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" Dec 05 10:53:47.395165 master-0 kubenswrapper[24928]: I1205 10:53:47.395091 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" Dec 05 10:53:47.395165 master-0 kubenswrapper[24928]: E1205 10:53:47.395112 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager" Dec 05 10:53:47.395165 master-0 kubenswrapper[24928]: I1205 10:53:47.395120 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager" Dec 05 10:53:47.395165 master-0 kubenswrapper[24928]: E1205 10:53:47.395133 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88cccb5b-1ad1-4fab-b34c-90252794ee20" containerName="route-controller-manager" Dec 05 10:53:47.395165 master-0 kubenswrapper[24928]: I1205 10:53:47.395141 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="88cccb5b-1ad1-4fab-b34c-90252794ee20" containerName="route-controller-manager" Dec 05 10:53:47.395165 master-0 kubenswrapper[24928]: E1205 10:53:47.395157 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager" Dec 05 10:53:47.395165 master-0 kubenswrapper[24928]: I1205 10:53:47.395165 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: E1205 10:53:47.395180 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88cccb5b-1ad1-4fab-b34c-90252794ee20" containerName="route-controller-manager" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: I1205 10:53:47.395190 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="88cccb5b-1ad1-4fab-b34c-90252794ee20" containerName="route-controller-manager" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: E1205 10:53:47.395208 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: I1205 10:53:47.395216 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: E1205 10:53:47.395226 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: I1205 10:53:47.395233 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: E1205 10:53:47.395247 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: I1205 10:53:47.395255 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: I1205 10:53:47.395388 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager-recovery-controller" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: I1205 10:53:47.395413 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="88cccb5b-1ad1-4fab-b34c-90252794ee20" containerName="route-controller-manager" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: I1205 10:53:47.395451 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager-recovery-controller" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: I1205 10:53:47.395465 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" Dec 05 10:53:47.395468 master-0 kubenswrapper[24928]: I1205 10:53:47.395475 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: I1205 10:53:47.395483 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: I1205 10:53:47.395491 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: I1205 10:53:47.395497 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: I1205 10:53:47.395508 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="88cccb5b-1ad1-4fab-b34c-90252794ee20" containerName="route-controller-manager" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: I1205 10:53:47.395517 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager-cert-syncer" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: I1205 10:53:47.395527 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: I1205 10:53:47.395544 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: I1205 10:53:47.395560 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: E1205 10:53:47.395698 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager-recovery-controller" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: I1205 10:53:47.395709 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager-recovery-controller" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: E1205 10:53:47.395725 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: I1205 10:53:47.395730 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="cluster-policy-controller" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: E1205 10:53:47.395746 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager" Dec 05 10:53:47.395830 master-0 kubenswrapper[24928]: I1205 10:53:47.395751 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager" Dec 05 10:53:47.396311 master-0 kubenswrapper[24928]: I1205 10:53:47.395854 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0cbad64-72b9-4ad3-9a42-4183e93c9ba0" containerName="controller-manager" Dec 05 10:53:47.396311 master-0 kubenswrapper[24928]: I1205 10:53:47.395873 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5219435a07a0220d41da97c4fb70abb1" containerName="kube-controller-manager" Dec 05 10:53:47.465333 master-0 kubenswrapper[24928]: I1205 10:53:47.465246 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f8788a6e4f698a66d5d8f663c8089807-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"f8788a6e4f698a66d5d8f663c8089807\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:47.465333 master-0 kubenswrapper[24928]: I1205 10:53:47.465319 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f8788a6e4f698a66d5d8f663c8089807-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"f8788a6e4f698a66d5d8f663c8089807\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:47.566318 master-0 kubenswrapper[24928]: I1205 10:53:47.566232 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f8788a6e4f698a66d5d8f663c8089807-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"f8788a6e4f698a66d5d8f663c8089807\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:47.566318 master-0 kubenswrapper[24928]: I1205 10:53:47.566302 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f8788a6e4f698a66d5d8f663c8089807-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"f8788a6e4f698a66d5d8f663c8089807\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:47.566634 master-0 kubenswrapper[24928]: I1205 10:53:47.566383 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f8788a6e4f698a66d5d8f663c8089807-resource-dir\") pod \"kube-controller-manager-master-0\" (UID: \"f8788a6e4f698a66d5d8f663c8089807\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:47.566634 master-0 kubenswrapper[24928]: I1205 10:53:47.566471 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f8788a6e4f698a66d5d8f663c8089807-cert-dir\") pod \"kube-controller-manager-master-0\" (UID: \"f8788a6e4f698a66d5d8f663c8089807\") " pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:47.589145 master-0 kubenswrapper[24928]: I1205 10:53:47.589093 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/cluster-policy-controller/1.log" Dec 05 10:53:47.590143 master-0 kubenswrapper[24928]: I1205 10:53:47.590081 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/1.log" Dec 05 10:53:47.591364 master-0 kubenswrapper[24928]: I1205 10:53:47.591124 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager-cert-syncer/0.log" Dec 05 10:53:47.591364 master-0 kubenswrapper[24928]: I1205 10:53:47.591220 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:47.594745 master-0 kubenswrapper[24928]: I1205 10:53:47.594686 24928 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" oldPodUID="5219435a07a0220d41da97c4fb70abb1" podUID="f8788a6e4f698a66d5d8f663c8089807" Dec 05 10:53:47.668085 master-0 kubenswrapper[24928]: I1205 10:53:47.667924 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-resource-dir\") pod \"5219435a07a0220d41da97c4fb70abb1\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " Dec 05 10:53:47.668270 master-0 kubenswrapper[24928]: I1205 10:53:47.668137 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-cert-dir\") pod \"5219435a07a0220d41da97c4fb70abb1\" (UID: \"5219435a07a0220d41da97c4fb70abb1\") " Dec 05 10:53:47.668309 master-0 kubenswrapper[24928]: I1205 10:53:47.668122 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "5219435a07a0220d41da97c4fb70abb1" (UID: "5219435a07a0220d41da97c4fb70abb1"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:53:47.668344 master-0 kubenswrapper[24928]: I1205 10:53:47.668232 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "5219435a07a0220d41da97c4fb70abb1" (UID: "5219435a07a0220d41da97c4fb70abb1"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:53:47.668539 master-0 kubenswrapper[24928]: I1205 10:53:47.668505 24928 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-cert-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:47.668539 master-0 kubenswrapper[24928]: I1205 10:53:47.668528 24928 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/5219435a07a0220d41da97c4fb70abb1-resource-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:48.045833 master-0 kubenswrapper[24928]: I1205 10:53:48.045736 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/cluster-policy-controller/1.log" Dec 05 10:53:48.046981 master-0 kubenswrapper[24928]: I1205 10:53:48.046933 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager/1.log" Dec 05 10:53:48.048267 master-0 kubenswrapper[24928]: I1205 10:53:48.048230 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager-cert-syncer/0.log" Dec 05 10:53:48.048371 master-0 kubenswrapper[24928]: I1205 10:53:48.048275 24928 generic.go:334] "Generic (PLEG): container finished" podID="5219435a07a0220d41da97c4fb70abb1" containerID="5e1cd037da3a72a86565cde7e25ecd2da3f7f8cf8ea4a83328e0ab5b87b966ad" exitCode=0 Dec 05 10:53:48.048371 master-0 kubenswrapper[24928]: I1205 10:53:48.048291 24928 generic.go:334] "Generic (PLEG): container finished" podID="5219435a07a0220d41da97c4fb70abb1" containerID="b27b1cff74cf1fa7645c9d670ba8742571627f768b820d46bb086bfbeb3ea29b" exitCode=0 Dec 05 10:53:48.048371 master-0 kubenswrapper[24928]: I1205 10:53:48.048300 24928 generic.go:334] "Generic (PLEG): container finished" podID="5219435a07a0220d41da97c4fb70abb1" containerID="50e85593f8bade43c99e607efe092eba239c2c489a561c3d4ee019a9a7b227cb" exitCode=0 Dec 05 10:53:48.048371 master-0 kubenswrapper[24928]: I1205 10:53:48.048358 24928 generic.go:334] "Generic (PLEG): container finished" podID="5219435a07a0220d41da97c4fb70abb1" containerID="47bef9607325414e871a594b06d4250139a33f93aa8b2368ce9838e31d093051" exitCode=2 Dec 05 10:53:48.048653 master-0 kubenswrapper[24928]: I1205 10:53:48.048437 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5be78d3569389b67ef6d6c6575e7c97a75617917f6e7f31b9b05a692b9e12292" Dec 05 10:53:48.048653 master-0 kubenswrapper[24928]: I1205 10:53:48.048468 24928 scope.go:117] "RemoveContainer" containerID="cf784a713d53bc0541dfbf03c915a4f6b70148e2c7b42b695d4ce6650a26f08f" Dec 05 10:53:48.049006 master-0 kubenswrapper[24928]: I1205 10:53:48.048893 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:48.050284 master-0 kubenswrapper[24928]: I1205 10:53:48.050253 24928 generic.go:334] "Generic (PLEG): container finished" podID="52a48f6c-b05f-4b70-96cc-378e37928c94" containerID="d539da331de2ee7e892ad095acecd0bf1191a2be47d514f0e5423eec137d52fd" exitCode=0 Dec 05 10:53:48.050359 master-0 kubenswrapper[24928]: I1205 10:53:48.050280 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-4-master-0" event={"ID":"52a48f6c-b05f-4b70-96cc-378e37928c94","Type":"ContainerDied","Data":"d539da331de2ee7e892ad095acecd0bf1191a2be47d514f0e5423eec137d52fd"} Dec 05 10:53:48.053483 master-0 kubenswrapper[24928]: I1205 10:53:48.053394 24928 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" oldPodUID="5219435a07a0220d41da97c4fb70abb1" podUID="f8788a6e4f698a66d5d8f663c8089807" Dec 05 10:53:48.077755 master-0 kubenswrapper[24928]: I1205 10:53:48.077689 24928 scope.go:117] "RemoveContainer" containerID="b81e8b68cc34c8fa308242ed9c31962b7b0b1e3af80c9d0154c5d9ae32a5982d" Dec 05 10:53:48.100668 master-0 kubenswrapper[24928]: I1205 10:53:48.100608 24928 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" oldPodUID="5219435a07a0220d41da97c4fb70abb1" podUID="f8788a6e4f698a66d5d8f663c8089807" Dec 05 10:53:48.111644 master-0 kubenswrapper[24928]: I1205 10:53:48.111584 24928 scope.go:117] "RemoveContainer" containerID="f7c86a3fbacaa184c777e4b5c80c29c00ccd9777b6aeed8a46d62880a41cc869" Dec 05 10:53:48.212300 master-0 kubenswrapper[24928]: I1205 10:53:48.212233 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5219435a07a0220d41da97c4fb70abb1" path="/var/lib/kubelet/pods/5219435a07a0220d41da97c4fb70abb1/volumes" Dec 05 10:53:49.060630 master-0 kubenswrapper[24928]: I1205 10:53:49.060564 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-master-0_5219435a07a0220d41da97c4fb70abb1/kube-controller-manager-cert-syncer/0.log" Dec 05 10:53:49.404016 master-0 kubenswrapper[24928]: I1205 10:53:49.403656 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:49.494684 master-0 kubenswrapper[24928]: I1205 10:53:49.493981 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52a48f6c-b05f-4b70-96cc-378e37928c94-kube-api-access\") pod \"52a48f6c-b05f-4b70-96cc-378e37928c94\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " Dec 05 10:53:49.494684 master-0 kubenswrapper[24928]: I1205 10:53:49.494086 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-kubelet-dir\") pod \"52a48f6c-b05f-4b70-96cc-378e37928c94\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " Dec 05 10:53:49.494684 master-0 kubenswrapper[24928]: I1205 10:53:49.494152 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-var-lock\") pod \"52a48f6c-b05f-4b70-96cc-378e37928c94\" (UID: \"52a48f6c-b05f-4b70-96cc-378e37928c94\") " Dec 05 10:53:49.494684 master-0 kubenswrapper[24928]: I1205 10:53:49.494222 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "52a48f6c-b05f-4b70-96cc-378e37928c94" (UID: "52a48f6c-b05f-4b70-96cc-378e37928c94"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:53:49.494684 master-0 kubenswrapper[24928]: I1205 10:53:49.494374 24928 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:49.494684 master-0 kubenswrapper[24928]: I1205 10:53:49.494368 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-var-lock" (OuterVolumeSpecName: "var-lock") pod "52a48f6c-b05f-4b70-96cc-378e37928c94" (UID: "52a48f6c-b05f-4b70-96cc-378e37928c94"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:53:49.497193 master-0 kubenswrapper[24928]: I1205 10:53:49.497100 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52a48f6c-b05f-4b70-96cc-378e37928c94-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "52a48f6c-b05f-4b70-96cc-378e37928c94" (UID: "52a48f6c-b05f-4b70-96cc-378e37928c94"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:53:49.595295 master-0 kubenswrapper[24928]: I1205 10:53:49.595233 24928 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/52a48f6c-b05f-4b70-96cc-378e37928c94-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:49.595295 master-0 kubenswrapper[24928]: I1205 10:53:49.595278 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/52a48f6c-b05f-4b70-96cc-378e37928c94-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:50.068735 master-0 kubenswrapper[24928]: I1205 10:53:50.068656 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/installer-4-master-0" event={"ID":"52a48f6c-b05f-4b70-96cc-378e37928c94","Type":"ContainerDied","Data":"83f85cf1dc2a5ddb90cf7d96dd9dbdd7fb07b05609bc0168ff762882c11c6a8c"} Dec 05 10:53:50.068735 master-0 kubenswrapper[24928]: I1205 10:53:50.068704 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83f85cf1dc2a5ddb90cf7d96dd9dbdd7fb07b05609bc0168ff762882c11c6a8c" Dec 05 10:53:50.069468 master-0 kubenswrapper[24928]: I1205 10:53:50.068823 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/installer-4-master-0" Dec 05 10:53:52.492544 master-0 kubenswrapper[24928]: I1205 10:53:52.492477 24928 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0"] Dec 05 10:53:52.493145 master-0 kubenswrapper[24928]: E1205 10:53:52.492794 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a48f6c-b05f-4b70-96cc-378e37928c94" containerName="installer" Dec 05 10:53:52.493145 master-0 kubenswrapper[24928]: I1205 10:53:52.492809 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a48f6c-b05f-4b70-96cc-378e37928c94" containerName="installer" Dec 05 10:53:52.493145 master-0 kubenswrapper[24928]: I1205 10:53:52.492969 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="52a48f6c-b05f-4b70-96cc-378e37928c94" containerName="installer" Dec 05 10:53:52.493436 master-0 kubenswrapper[24928]: I1205 10:53:52.493373 24928 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-master-0"] Dec 05 10:53:52.493741 master-0 kubenswrapper[24928]: I1205 10:53:52.493695 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5" gracePeriod=15 Dec 05 10:53:52.493820 master-0 kubenswrapper[24928]: I1205 10:53:52.493745 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.493879 master-0 kubenswrapper[24928]: I1205 10:53:52.493826 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-check-endpoints" containerID="cri-o://eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5" gracePeriod=15 Dec 05 10:53:52.493879 master-0 kubenswrapper[24928]: I1205 10:53:52.493677 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver" containerID="cri-o://a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56" gracePeriod=15 Dec 05 10:53:52.493879 master-0 kubenswrapper[24928]: I1205 10:53:52.493803 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-cert-syncer" containerID="cri-o://7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63" gracePeriod=15 Dec 05 10:53:52.494051 master-0 kubenswrapper[24928]: I1205 10:53:52.493884 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf" gracePeriod=15 Dec 05 10:53:52.495066 master-0 kubenswrapper[24928]: I1205 10:53:52.495008 24928 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-master-0"] Dec 05 10:53:52.495402 master-0 kubenswrapper[24928]: E1205 10:53:52.495372 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver" Dec 05 10:53:52.495504 master-0 kubenswrapper[24928]: I1205 10:53:52.495403 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver" Dec 05 10:53:52.495504 master-0 kubenswrapper[24928]: E1205 10:53:52.495461 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-cert-syncer" Dec 05 10:53:52.495504 master-0 kubenswrapper[24928]: I1205 10:53:52.495481 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-cert-syncer" Dec 05 10:53:52.495641 master-0 kubenswrapper[24928]: E1205 10:53:52.495514 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 10:53:52.495641 master-0 kubenswrapper[24928]: I1205 10:53:52.495532 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 10:53:52.495641 master-0 kubenswrapper[24928]: E1205 10:53:52.495552 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-check-endpoints" Dec 05 10:53:52.495641 master-0 kubenswrapper[24928]: I1205 10:53:52.495565 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-check-endpoints" Dec 05 10:53:52.495641 master-0 kubenswrapper[24928]: E1205 10:53:52.495598 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-insecure-readyz" Dec 05 10:53:52.495641 master-0 kubenswrapper[24928]: I1205 10:53:52.495611 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-insecure-readyz" Dec 05 10:53:52.495943 master-0 kubenswrapper[24928]: E1205 10:53:52.495646 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="setup" Dec 05 10:53:52.495943 master-0 kubenswrapper[24928]: I1205 10:53:52.495660 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="setup" Dec 05 10:53:52.495943 master-0 kubenswrapper[24928]: I1205 10:53:52.495904 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-insecure-readyz" Dec 05 10:53:52.495943 master-0 kubenswrapper[24928]: I1205 10:53:52.495924 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="setup" Dec 05 10:53:52.496143 master-0 kubenswrapper[24928]: I1205 10:53:52.495959 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver" Dec 05 10:53:52.496143 master-0 kubenswrapper[24928]: I1205 10:53:52.495973 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-check-endpoints" Dec 05 10:53:52.496143 master-0 kubenswrapper[24928]: I1205 10:53:52.495997 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 10:53:52.496143 master-0 kubenswrapper[24928]: I1205 10:53:52.496030 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="b89698aa356a3bc32694e2b098f9a900" containerName="kube-apiserver-cert-syncer" Dec 05 10:53:52.537969 master-0 kubenswrapper[24928]: I1205 10:53:52.537847 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.537969 master-0 kubenswrapper[24928]: I1205 10:53:52.537976 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a369cadf0161d66f2936cdea3ded59b7-audit-dir\") pod \"kube-apiserver-master-0\" (UID: \"a369cadf0161d66f2936cdea3ded59b7\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:53:52.538250 master-0 kubenswrapper[24928]: I1205 10:53:52.538023 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a369cadf0161d66f2936cdea3ded59b7-resource-dir\") pod \"kube-apiserver-master-0\" (UID: \"a369cadf0161d66f2936cdea3ded59b7\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:53:52.538250 master-0 kubenswrapper[24928]: I1205 10:53:52.538041 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-log\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.538250 master-0 kubenswrapper[24928]: I1205 10:53:52.538090 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.538250 master-0 kubenswrapper[24928]: I1205 10:53:52.538110 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-lock\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.538250 master-0 kubenswrapper[24928]: I1205 10:53:52.538142 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/a369cadf0161d66f2936cdea3ded59b7-cert-dir\") pod \"kube-apiserver-master-0\" (UID: \"a369cadf0161d66f2936cdea3ded59b7\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:53:52.538250 master-0 kubenswrapper[24928]: I1205 10:53:52.538168 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-manifests\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.562553 master-0 kubenswrapper[24928]: I1205 10:53:52.560868 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0"] Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.639948 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640017 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-lock\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640109 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-lock\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640155 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/a369cadf0161d66f2936cdea3ded59b7-cert-dir\") pod \"kube-apiserver-master-0\" (UID: \"a369cadf0161d66f2936cdea3ded59b7\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640179 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-manifests\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640198 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640231 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640258 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-manifests\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640212 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-resource-dir\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640274 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/a369cadf0161d66f2936cdea3ded59b7-cert-dir\") pod \"kube-apiserver-master-0\" (UID: \"a369cadf0161d66f2936cdea3ded59b7\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640297 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a369cadf0161d66f2936cdea3ded59b7-audit-dir\") pod \"kube-apiserver-master-0\" (UID: \"a369cadf0161d66f2936cdea3ded59b7\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640281 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a369cadf0161d66f2936cdea3ded59b7-audit-dir\") pod \"kube-apiserver-master-0\" (UID: \"a369cadf0161d66f2936cdea3ded59b7\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640335 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a369cadf0161d66f2936cdea3ded59b7-resource-dir\") pod \"kube-apiserver-master-0\" (UID: \"a369cadf0161d66f2936cdea3ded59b7\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640352 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-log\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640414 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-log\") pod \"kube-apiserver-startup-monitor-master-0\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.647740 master-0 kubenswrapper[24928]: I1205 10:53:52.640455 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/a369cadf0161d66f2936cdea3ded59b7-resource-dir\") pod \"kube-apiserver-master-0\" (UID: \"a369cadf0161d66f2936cdea3ded59b7\") " pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:53:52.858284 master-0 kubenswrapper[24928]: I1205 10:53:52.858182 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:53:52.880044 master-0 kubenswrapper[24928]: W1205 10:53:52.879969 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb83ccd6fa217a93a2c607d0109896ef8.slice/crio-80a6d9cf544606b0c1f9ca1366aa017d15145b4a3a07d7a47adb7ed220305368 WatchSource:0}: Error finding container 80a6d9cf544606b0c1f9ca1366aa017d15145b4a3a07d7a47adb7ed220305368: Status 404 returned error can't find the container with id 80a6d9cf544606b0c1f9ca1366aa017d15145b4a3a07d7a47adb7ed220305368 Dec 05 10:53:52.883593 master-0 kubenswrapper[24928]: E1205 10:53:52.883467 24928 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 192.168.32.10:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-master-0.187e4c5538a98dff openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-master-0,UID:b83ccd6fa217a93a2c607d0109896ef8,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660\" already present on machine,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:53:52.882380287 +0000 UTC m=+392.885574138,LastTimestamp:2025-12-05 10:53:52.882380287 +0000 UTC m=+392.885574138,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:53:53.094683 master-0 kubenswrapper[24928]: I1205 10:53:53.094616 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-master-0_b89698aa356a3bc32694e2b098f9a900/kube-apiserver-cert-syncer/0.log" Dec 05 10:53:53.095722 master-0 kubenswrapper[24928]: I1205 10:53:53.095623 24928 generic.go:334] "Generic (PLEG): container finished" podID="b89698aa356a3bc32694e2b098f9a900" containerID="eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5" exitCode=0 Dec 05 10:53:53.095865 master-0 kubenswrapper[24928]: I1205 10:53:53.095721 24928 generic.go:334] "Generic (PLEG): container finished" podID="b89698aa356a3bc32694e2b098f9a900" containerID="8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5" exitCode=0 Dec 05 10:53:53.095865 master-0 kubenswrapper[24928]: I1205 10:53:53.095742 24928 generic.go:334] "Generic (PLEG): container finished" podID="b89698aa356a3bc32694e2b098f9a900" containerID="9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf" exitCode=0 Dec 05 10:53:53.095865 master-0 kubenswrapper[24928]: I1205 10:53:53.095756 24928 generic.go:334] "Generic (PLEG): container finished" podID="b89698aa356a3bc32694e2b098f9a900" containerID="7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63" exitCode=2 Dec 05 10:53:53.097357 master-0 kubenswrapper[24928]: I1205 10:53:53.097300 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" event={"ID":"b83ccd6fa217a93a2c607d0109896ef8","Type":"ContainerStarted","Data":"80a6d9cf544606b0c1f9ca1366aa017d15145b4a3a07d7a47adb7ed220305368"} Dec 05 10:53:53.099023 master-0 kubenswrapper[24928]: I1205 10:53:53.098969 24928 generic.go:334] "Generic (PLEG): container finished" podID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" containerID="8e36c34ac1ad0faf4333c95bccd24ab7f0d2a8d130aee4d157dbd04cb3cc5423" exitCode=0 Dec 05 10:53:53.099023 master-0 kubenswrapper[24928]: I1205 10:53:53.099012 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-5-master-0" event={"ID":"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0","Type":"ContainerDied","Data":"8e36c34ac1ad0faf4333c95bccd24ab7f0d2a8d130aee4d157dbd04cb3cc5423"} Dec 05 10:53:53.101965 master-0 kubenswrapper[24928]: I1205 10:53:53.101914 24928 status_manager.go:851] "Failed to get status for pod" podUID="b89698aa356a3bc32694e2b098f9a900" pod="openshift-kube-apiserver/kube-apiserver-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:53.102872 master-0 kubenswrapper[24928]: I1205 10:53:53.102782 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:53.103741 master-0 kubenswrapper[24928]: I1205 10:53:53.103661 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:54.106415 master-0 kubenswrapper[24928]: I1205 10:53:54.106318 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" event={"ID":"b83ccd6fa217a93a2c607d0109896ef8","Type":"ContainerStarted","Data":"b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38"} Dec 05 10:53:54.108084 master-0 kubenswrapper[24928]: I1205 10:53:54.108004 24928 status_manager.go:851] "Failed to get status for pod" podUID="b89698aa356a3bc32694e2b098f9a900" pod="openshift-kube-apiserver/kube-apiserver-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:54.108626 master-0 kubenswrapper[24928]: I1205 10:53:54.108580 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:54.109253 master-0 kubenswrapper[24928]: I1205 10:53:54.109176 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:54.475581 master-0 kubenswrapper[24928]: I1205 10:53:54.475219 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:54.476222 master-0 kubenswrapper[24928]: I1205 10:53:54.476168 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:54.476730 master-0 kubenswrapper[24928]: I1205 10:53:54.476687 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:54.578121 master-0 kubenswrapper[24928]: I1205 10:53:54.578034 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kube-api-access\") pod \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " Dec 05 10:53:54.578121 master-0 kubenswrapper[24928]: I1205 10:53:54.578130 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-var-lock\") pod \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " Dec 05 10:53:54.578121 master-0 kubenswrapper[24928]: I1205 10:53:54.578162 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kubelet-dir\") pod \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\" (UID: \"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0\") " Dec 05 10:53:54.578890 master-0 kubenswrapper[24928]: I1205 10:53:54.578268 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-var-lock" (OuterVolumeSpecName: "var-lock") pod "4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" (UID: "4d237cdd-9aa8-4a12-b244-1dc1b4c456d0"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:53:54.578890 master-0 kubenswrapper[24928]: I1205 10:53:54.578373 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" (UID: "4d237cdd-9aa8-4a12-b244-1dc1b4c456d0"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:53:54.578890 master-0 kubenswrapper[24928]: I1205 10:53:54.578606 24928 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:54.578890 master-0 kubenswrapper[24928]: I1205 10:53:54.578636 24928 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kubelet-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:54.580895 master-0 kubenswrapper[24928]: I1205 10:53:54.580813 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" (UID: "4d237cdd-9aa8-4a12-b244-1dc1b4c456d0"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:53:54.680361 master-0 kubenswrapper[24928]: I1205 10:53:54.680303 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4d237cdd-9aa8-4a12-b244-1dc1b4c456d0-kube-api-access\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:54.883783 master-0 kubenswrapper[24928]: I1205 10:53:54.883751 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-master-0_b89698aa356a3bc32694e2b098f9a900/kube-apiserver-cert-syncer/0.log" Dec 05 10:53:54.884455 master-0 kubenswrapper[24928]: I1205 10:53:54.884409 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:53:54.885478 master-0 kubenswrapper[24928]: I1205 10:53:54.885405 24928 status_manager.go:851] "Failed to get status for pod" podUID="b89698aa356a3bc32694e2b098f9a900" pod="openshift-kube-apiserver/kube-apiserver-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:54.886011 master-0 kubenswrapper[24928]: I1205 10:53:54.885975 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:54.886582 master-0 kubenswrapper[24928]: I1205 10:53:54.886544 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:54.903853 master-0 kubenswrapper[24928]: I1205 10:53:54.903819 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:53:54.904005 master-0 kubenswrapper[24928]: I1205 10:53:54.903860 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:53:54.983679 master-0 kubenswrapper[24928]: I1205 10:53:54.983594 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-cert-dir\") pod \"b89698aa356a3bc32694e2b098f9a900\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " Dec 05 10:53:54.983977 master-0 kubenswrapper[24928]: I1205 10:53:54.983743 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "b89698aa356a3bc32694e2b098f9a900" (UID: "b89698aa356a3bc32694e2b098f9a900"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:53:54.983977 master-0 kubenswrapper[24928]: I1205 10:53:54.983833 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-resource-dir\") pod \"b89698aa356a3bc32694e2b098f9a900\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " Dec 05 10:53:54.983977 master-0 kubenswrapper[24928]: I1205 10:53:54.983911 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-audit-dir\") pod \"b89698aa356a3bc32694e2b098f9a900\" (UID: \"b89698aa356a3bc32694e2b098f9a900\") " Dec 05 10:53:54.983977 master-0 kubenswrapper[24928]: I1205 10:53:54.983957 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "b89698aa356a3bc32694e2b098f9a900" (UID: "b89698aa356a3bc32694e2b098f9a900"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:53:54.984244 master-0 kubenswrapper[24928]: I1205 10:53:54.984077 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "b89698aa356a3bc32694e2b098f9a900" (UID: "b89698aa356a3bc32694e2b098f9a900"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:53:54.984401 master-0 kubenswrapper[24928]: I1205 10:53:54.984373 24928 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-resource-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:54.984496 master-0 kubenswrapper[24928]: I1205 10:53:54.984444 24928 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-audit-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:54.984496 master-0 kubenswrapper[24928]: I1205 10:53:54.984465 24928 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/b89698aa356a3bc32694e2b098f9a900-cert-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:55.116873 master-0 kubenswrapper[24928]: I1205 10:53:55.116825 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-5-master-0" Dec 05 10:53:55.117468 master-0 kubenswrapper[24928]: I1205 10:53:55.116814 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-5-master-0" event={"ID":"4d237cdd-9aa8-4a12-b244-1dc1b4c456d0","Type":"ContainerDied","Data":"24059d1d31317d09c58db0a6c95ce43a0078bec9901eb497277dd670e3b3751e"} Dec 05 10:53:55.117468 master-0 kubenswrapper[24928]: I1205 10:53:55.117162 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24059d1d31317d09c58db0a6c95ce43a0078bec9901eb497277dd670e3b3751e" Dec 05 10:53:55.119972 master-0 kubenswrapper[24928]: I1205 10:53:55.119923 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-master-0_b89698aa356a3bc32694e2b098f9a900/kube-apiserver-cert-syncer/0.log" Dec 05 10:53:55.120681 master-0 kubenswrapper[24928]: I1205 10:53:55.120643 24928 generic.go:334] "Generic (PLEG): container finished" podID="b89698aa356a3bc32694e2b098f9a900" containerID="a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56" exitCode=0 Dec 05 10:53:55.120759 master-0 kubenswrapper[24928]: I1205 10:53:55.120727 24928 scope.go:117] "RemoveContainer" containerID="eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5" Dec 05 10:53:55.120816 master-0 kubenswrapper[24928]: I1205 10:53:55.120759 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:53:55.143909 master-0 kubenswrapper[24928]: I1205 10:53:55.143849 24928 scope.go:117] "RemoveContainer" containerID="8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5" Dec 05 10:53:55.156258 master-0 kubenswrapper[24928]: I1205 10:53:55.156166 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:55.157161 master-0 kubenswrapper[24928]: I1205 10:53:55.157070 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:55.158002 master-0 kubenswrapper[24928]: I1205 10:53:55.157935 24928 status_manager.go:851] "Failed to get status for pod" podUID="b89698aa356a3bc32694e2b098f9a900" pod="openshift-kube-apiserver/kube-apiserver-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:55.158788 master-0 kubenswrapper[24928]: I1205 10:53:55.158723 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:55.159311 master-0 kubenswrapper[24928]: I1205 10:53:55.159246 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:55.159911 master-0 kubenswrapper[24928]: I1205 10:53:55.159856 24928 status_manager.go:851] "Failed to get status for pod" podUID="b89698aa356a3bc32694e2b098f9a900" pod="openshift-kube-apiserver/kube-apiserver-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:55.176033 master-0 kubenswrapper[24928]: I1205 10:53:55.175962 24928 scope.go:117] "RemoveContainer" containerID="9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf" Dec 05 10:53:55.198247 master-0 kubenswrapper[24928]: I1205 10:53:55.198171 24928 scope.go:117] "RemoveContainer" containerID="7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63" Dec 05 10:53:55.215126 master-0 kubenswrapper[24928]: I1205 10:53:55.215069 24928 scope.go:117] "RemoveContainer" containerID="a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56" Dec 05 10:53:55.233300 master-0 kubenswrapper[24928]: I1205 10:53:55.233250 24928 scope.go:117] "RemoveContainer" containerID="c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b" Dec 05 10:53:55.255831 master-0 kubenswrapper[24928]: I1205 10:53:55.255778 24928 scope.go:117] "RemoveContainer" containerID="eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5" Dec 05 10:53:55.256554 master-0 kubenswrapper[24928]: E1205 10:53:55.256495 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5\": container with ID starting with eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5 not found: ID does not exist" containerID="eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5" Dec 05 10:53:55.256620 master-0 kubenswrapper[24928]: I1205 10:53:55.256568 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5"} err="failed to get container status \"eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5\": rpc error: code = NotFound desc = could not find container \"eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5\": container with ID starting with eb38c2721da5b76773ce9f360b2f381759e040cae4be30efa649ae34ff1f70a5 not found: ID does not exist" Dec 05 10:53:55.256662 master-0 kubenswrapper[24928]: I1205 10:53:55.256618 24928 scope.go:117] "RemoveContainer" containerID="8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5" Dec 05 10:53:55.257156 master-0 kubenswrapper[24928]: E1205 10:53:55.257118 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5\": container with ID starting with 8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5 not found: ID does not exist" containerID="8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5" Dec 05 10:53:55.257195 master-0 kubenswrapper[24928]: I1205 10:53:55.257153 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5"} err="failed to get container status \"8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5\": rpc error: code = NotFound desc = could not find container \"8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5\": container with ID starting with 8a7e16ffba6dbe189e8b6323a0a8253c4aa4135ef0dc2efd0edde83630330be5 not found: ID does not exist" Dec 05 10:53:55.257195 master-0 kubenswrapper[24928]: I1205 10:53:55.257174 24928 scope.go:117] "RemoveContainer" containerID="9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf" Dec 05 10:53:55.257602 master-0 kubenswrapper[24928]: E1205 10:53:55.257565 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf\": container with ID starting with 9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf not found: ID does not exist" containerID="9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf" Dec 05 10:53:55.257602 master-0 kubenswrapper[24928]: I1205 10:53:55.257588 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf"} err="failed to get container status \"9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf\": rpc error: code = NotFound desc = could not find container \"9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf\": container with ID starting with 9c66bfdc5f429381c3516a36dfd76d4a6d909377e8884734cfb6a5586e69e3bf not found: ID does not exist" Dec 05 10:53:55.257602 master-0 kubenswrapper[24928]: I1205 10:53:55.257600 24928 scope.go:117] "RemoveContainer" containerID="7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63" Dec 05 10:53:55.257985 master-0 kubenswrapper[24928]: E1205 10:53:55.257946 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63\": container with ID starting with 7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63 not found: ID does not exist" containerID="7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63" Dec 05 10:53:55.257985 master-0 kubenswrapper[24928]: I1205 10:53:55.257974 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63"} err="failed to get container status \"7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63\": rpc error: code = NotFound desc = could not find container \"7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63\": container with ID starting with 7774b8bac576563565ea420e846478b215157027cfe3803f551651976387fd63 not found: ID does not exist" Dec 05 10:53:55.258073 master-0 kubenswrapper[24928]: I1205 10:53:55.257990 24928 scope.go:117] "RemoveContainer" containerID="a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56" Dec 05 10:53:55.258331 master-0 kubenswrapper[24928]: E1205 10:53:55.258284 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56\": container with ID starting with a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56 not found: ID does not exist" containerID="a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56" Dec 05 10:53:55.258381 master-0 kubenswrapper[24928]: I1205 10:53:55.258330 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56"} err="failed to get container status \"a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56\": rpc error: code = NotFound desc = could not find container \"a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56\": container with ID starting with a832b5702638b7d0cb407c271f68a16cc9bc8b47363b8669f5ad6bf78d036c56 not found: ID does not exist" Dec 05 10:53:55.258381 master-0 kubenswrapper[24928]: I1205 10:53:55.258375 24928 scope.go:117] "RemoveContainer" containerID="c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b" Dec 05 10:53:55.258810 master-0 kubenswrapper[24928]: E1205 10:53:55.258767 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b\": container with ID starting with c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b not found: ID does not exist" containerID="c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b" Dec 05 10:53:55.258867 master-0 kubenswrapper[24928]: I1205 10:53:55.258803 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b"} err="failed to get container status \"c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b\": rpc error: code = NotFound desc = could not find container \"c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b\": container with ID starting with c20621bfc1ac53a74ca8919137809b54777813f204fa43a533fde001fd704d3b not found: ID does not exist" Dec 05 10:53:56.212275 master-0 kubenswrapper[24928]: I1205 10:53:56.212221 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b89698aa356a3bc32694e2b098f9a900" path="/var/lib/kubelet/pods/b89698aa356a3bc32694e2b098f9a900/volumes" Dec 05 10:53:57.068807 master-0 kubenswrapper[24928]: E1205 10:53:57.068567 24928 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 192.168.32.10:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-master-0.187e4c5538a98dff openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-master-0,UID:b83ccd6fa217a93a2c607d0109896ef8,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660\" already present on machine,Source:EventSource{Component:kubelet,Host:master-0,},FirstTimestamp:2025-12-05 10:53:52.882380287 +0000 UTC m=+392.885574138,LastTimestamp:2025-12-05 10:53:52.882380287 +0000 UTC m=+392.885574138,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:master-0,}" Dec 05 10:53:59.203160 master-0 kubenswrapper[24928]: I1205 10:53:59.203069 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-74f96dcf4d-9gskd" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" containerID="cri-o://ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311" gracePeriod=15 Dec 05 10:53:59.203790 master-0 kubenswrapper[24928]: I1205 10:53:59.203649 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:59.204211 master-0 kubenswrapper[24928]: I1205 10:53:59.204165 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:53:59.204211 master-0 kubenswrapper[24928]: I1205 10:53:59.204209 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:53:59.205147 master-0 kubenswrapper[24928]: I1205 10:53:59.205063 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:59.205147 master-0 kubenswrapper[24928]: E1205 10:53:59.205114 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-etcd/pods/etcd-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" pod="openshift-etcd/etcd-master-0" Dec 05 10:53:59.206194 master-0 kubenswrapper[24928]: I1205 10:53:59.206142 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:59.308371 master-0 kubenswrapper[24928]: I1205 10:53:59.308280 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:53:59.308371 master-0 kubenswrapper[24928]: I1205 10:53:59.308339 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:53:59.309241 master-0 kubenswrapper[24928]: E1205 10:53:59.309180 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:59.309896 master-0 kubenswrapper[24928]: I1205 10:53:59.309847 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:53:59.345470 master-0 kubenswrapper[24928]: W1205 10:53:59.345375 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8788a6e4f698a66d5d8f663c8089807.slice/crio-e916e5cf6d62b88495f4029cb7b85e45148cab71a499aae3b999683e64a1b8c2 WatchSource:0}: Error finding container e916e5cf6d62b88495f4029cb7b85e45148cab71a499aae3b999683e64a1b8c2: Status 404 returned error can't find the container with id e916e5cf6d62b88495f4029cb7b85e45148cab71a499aae3b999683e64a1b8c2 Dec 05 10:53:59.698566 master-0 kubenswrapper[24928]: I1205 10:53:59.698506 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74f96dcf4d-9gskd_3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527/console/1.log" Dec 05 10:53:59.699196 master-0 kubenswrapper[24928]: I1205 10:53:59.699146 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74f96dcf4d-9gskd_3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527/console/0.log" Dec 05 10:53:59.699287 master-0 kubenswrapper[24928]: I1205 10:53:59.699238 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:53:59.700396 master-0 kubenswrapper[24928]: I1205 10:53:59.700338 24928 status_manager.go:851] "Failed to get status for pod" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" pod="openshift-console/console-74f96dcf4d-9gskd" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-console/pods/console-74f96dcf4d-9gskd\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:59.701381 master-0 kubenswrapper[24928]: I1205 10:53:59.701346 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:59.703009 master-0 kubenswrapper[24928]: I1205 10:53:59.702959 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:53:59.752267 master-0 kubenswrapper[24928]: I1205 10:53:59.752175 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-service-ca\") pod \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " Dec 05 10:53:59.752267 master-0 kubenswrapper[24928]: I1205 10:53:59.752258 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-oauth-serving-cert\") pod \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " Dec 05 10:53:59.752515 master-0 kubenswrapper[24928]: I1205 10:53:59.752354 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsgbq\" (UniqueName: \"kubernetes.io/projected/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-kube-api-access-nsgbq\") pod \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " Dec 05 10:53:59.752515 master-0 kubenswrapper[24928]: I1205 10:53:59.752381 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-config\") pod \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " Dec 05 10:53:59.752515 master-0 kubenswrapper[24928]: I1205 10:53:59.752440 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-serving-cert\") pod \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " Dec 05 10:53:59.752605 master-0 kubenswrapper[24928]: I1205 10:53:59.752514 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-oauth-config\") pod \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\" (UID: \"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527\") " Dec 05 10:53:59.753338 master-0 kubenswrapper[24928]: I1205 10:53:59.753281 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-service-ca" (OuterVolumeSpecName: "service-ca") pod "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" (UID: "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:53:59.753927 master-0 kubenswrapper[24928]: I1205 10:53:59.753897 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" (UID: "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:53:59.754006 master-0 kubenswrapper[24928]: I1205 10:53:59.753890 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-config" (OuterVolumeSpecName: "console-config") pod "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" (UID: "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:53:59.756160 master-0 kubenswrapper[24928]: I1205 10:53:59.756117 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" (UID: "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:53:59.757158 master-0 kubenswrapper[24928]: I1205 10:53:59.756686 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" (UID: "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:53:59.758149 master-0 kubenswrapper[24928]: I1205 10:53:59.758074 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-kube-api-access-nsgbq" (OuterVolumeSpecName: "kube-api-access-nsgbq") pod "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" (UID: "3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527"). InnerVolumeSpecName "kube-api-access-nsgbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:53:59.853735 master-0 kubenswrapper[24928]: I1205 10:53:59.853646 24928 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-oauth-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:59.853735 master-0 kubenswrapper[24928]: I1205 10:53:59.853699 24928 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-service-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:59.853735 master-0 kubenswrapper[24928]: I1205 10:53:59.853710 24928 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-oauth-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:59.853735 master-0 kubenswrapper[24928]: I1205 10:53:59.853725 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsgbq\" (UniqueName: \"kubernetes.io/projected/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-kube-api-access-nsgbq\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:59.853735 master-0 kubenswrapper[24928]: I1205 10:53:59.853735 24928 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:53:59.853735 master-0 kubenswrapper[24928]: I1205 10:53:59.853745 24928 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527-console-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:54:00.149358 master-0 kubenswrapper[24928]: E1205 10:54:00.149287 24928 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.150316 master-0 kubenswrapper[24928]: E1205 10:54:00.150249 24928 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.151237 master-0 kubenswrapper[24928]: E1205 10:54:00.151145 24928 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.152144 master-0 kubenswrapper[24928]: E1205 10:54:00.152076 24928 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.152841 master-0 kubenswrapper[24928]: E1205 10:54:00.152793 24928 controller.go:195] "Failed to update lease" err="Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.152841 master-0 kubenswrapper[24928]: I1205 10:54:00.152836 24928 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 05 10:54:00.153739 master-0 kubenswrapper[24928]: E1205 10:54:00.153638 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="200ms" Dec 05 10:54:00.165214 master-0 kubenswrapper[24928]: I1205 10:54:00.165142 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"f8788a6e4f698a66d5d8f663c8089807","Type":"ContainerStarted","Data":"07105d866ba818492f7da826a5047a599cc56caa1b70d9b79cc54ab46b275a10"} Dec 05 10:54:00.165407 master-0 kubenswrapper[24928]: I1205 10:54:00.165229 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"f8788a6e4f698a66d5d8f663c8089807","Type":"ContainerStarted","Data":"df18ed60acc8bd1cf2ac3478a7dabb2cff45ea28b2fb9081190f1b4db8f3e6b3"} Dec 05 10:54:00.165407 master-0 kubenswrapper[24928]: I1205 10:54:00.165259 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"f8788a6e4f698a66d5d8f663c8089807","Type":"ContainerStarted","Data":"e916e5cf6d62b88495f4029cb7b85e45148cab71a499aae3b999683e64a1b8c2"} Dec 05 10:54:00.167853 master-0 kubenswrapper[24928]: I1205 10:54:00.167791 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74f96dcf4d-9gskd_3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527/console/1.log" Dec 05 10:54:00.168858 master-0 kubenswrapper[24928]: I1205 10:54:00.168724 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74f96dcf4d-9gskd_3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527/console/0.log" Dec 05 10:54:00.168858 master-0 kubenswrapper[24928]: I1205 10:54:00.168794 24928 generic.go:334] "Generic (PLEG): container finished" podID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerID="ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311" exitCode=2 Dec 05 10:54:00.169041 master-0 kubenswrapper[24928]: I1205 10:54:00.168849 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74f96dcf4d-9gskd" event={"ID":"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527","Type":"ContainerDied","Data":"ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311"} Dec 05 10:54:00.169041 master-0 kubenswrapper[24928]: I1205 10:54:00.168894 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74f96dcf4d-9gskd" Dec 05 10:54:00.169041 master-0 kubenswrapper[24928]: I1205 10:54:00.168934 24928 scope.go:117] "RemoveContainer" containerID="ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311" Dec 05 10:54:00.169041 master-0 kubenswrapper[24928]: I1205 10:54:00.168917 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74f96dcf4d-9gskd" event={"ID":"3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527","Type":"ContainerDied","Data":"940d3650e9f9322807a81f39b2daf6291f7767c05f0fcc28414d6332cac43bcd"} Dec 05 10:54:00.174987 master-0 kubenswrapper[24928]: I1205 10:54:00.174924 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.176474 master-0 kubenswrapper[24928]: I1205 10:54:00.176353 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.178908 master-0 kubenswrapper[24928]: I1205 10:54:00.178834 24928 status_manager.go:851] "Failed to get status for pod" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" pod="openshift-console/console-74f96dcf4d-9gskd" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-console/pods/console-74f96dcf4d-9gskd\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.189979 master-0 kubenswrapper[24928]: I1205 10:54:00.189893 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.190658 master-0 kubenswrapper[24928]: I1205 10:54:00.190590 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.191563 master-0 kubenswrapper[24928]: I1205 10:54:00.191486 24928 status_manager.go:851] "Failed to get status for pod" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" pod="openshift-console/console-74f96dcf4d-9gskd" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-console/pods/console-74f96dcf4d-9gskd\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.209460 master-0 kubenswrapper[24928]: I1205 10:54:00.209346 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.211116 master-0 kubenswrapper[24928]: I1205 10:54:00.210689 24928 status_manager.go:851] "Failed to get status for pod" podUID="f8788a6e4f698a66d5d8f663c8089807" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.211116 master-0 kubenswrapper[24928]: I1205 10:54:00.210858 24928 scope.go:117] "RemoveContainer" containerID="61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2" Dec 05 10:54:00.212103 master-0 kubenswrapper[24928]: I1205 10:54:00.212047 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.213113 master-0 kubenswrapper[24928]: I1205 10:54:00.213062 24928 status_manager.go:851] "Failed to get status for pod" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" pod="openshift-console/console-74f96dcf4d-9gskd" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-console/pods/console-74f96dcf4d-9gskd\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:00.234023 master-0 kubenswrapper[24928]: I1205 10:54:00.233965 24928 scope.go:117] "RemoveContainer" containerID="ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311" Dec 05 10:54:00.234601 master-0 kubenswrapper[24928]: E1205 10:54:00.234542 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311\": container with ID starting with ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311 not found: ID does not exist" containerID="ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311" Dec 05 10:54:00.234702 master-0 kubenswrapper[24928]: I1205 10:54:00.234595 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311"} err="failed to get container status \"ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311\": rpc error: code = NotFound desc = could not find container \"ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311\": container with ID starting with ad8ebfeea9223543660d22babd6e5b8fbeec2a35cae900225e50512add0ab311 not found: ID does not exist" Dec 05 10:54:00.234702 master-0 kubenswrapper[24928]: I1205 10:54:00.234621 24928 scope.go:117] "RemoveContainer" containerID="61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2" Dec 05 10:54:00.235108 master-0 kubenswrapper[24928]: E1205 10:54:00.235068 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2\": container with ID starting with 61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2 not found: ID does not exist" containerID="61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2" Dec 05 10:54:00.235108 master-0 kubenswrapper[24928]: I1205 10:54:00.235090 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2"} err="failed to get container status \"61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2\": rpc error: code = NotFound desc = could not find container \"61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2\": container with ID starting with 61f6cd0449ddc0d9833473262f880e6dede90eeb835995174d194502183a98a2 not found: ID does not exist" Dec 05 10:54:00.355328 master-0 kubenswrapper[24928]: E1205 10:54:00.355219 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="400ms" Dec 05 10:54:00.756878 master-0 kubenswrapper[24928]: E1205 10:54:00.756674 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="800ms" Dec 05 10:54:01.182196 master-0 kubenswrapper[24928]: I1205 10:54:01.182124 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"f8788a6e4f698a66d5d8f663c8089807","Type":"ContainerStarted","Data":"2ee8b14ae8fd0662c54c4bb05d795ea164341bef136560454e443fbf2e7f141e"} Dec 05 10:54:01.182196 master-0 kubenswrapper[24928]: I1205 10:54:01.182186 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" event={"ID":"f8788a6e4f698a66d5d8f663c8089807","Type":"ContainerStarted","Data":"3a29d100912f8f7afbda2a999da94e22fecef3a43100a5f96cc5f72f261db86c"} Dec 05 10:54:01.182494 master-0 kubenswrapper[24928]: I1205 10:54:01.182370 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:01.182494 master-0 kubenswrapper[24928]: I1205 10:54:01.182402 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:01.183400 master-0 kubenswrapper[24928]: I1205 10:54:01.183340 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:01.183491 master-0 kubenswrapper[24928]: E1205 10:54:01.183387 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:54:01.184206 master-0 kubenswrapper[24928]: I1205 10:54:01.184142 24928 status_manager.go:851] "Failed to get status for pod" podUID="f8788a6e4f698a66d5d8f663c8089807" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:01.184787 master-0 kubenswrapper[24928]: I1205 10:54:01.184745 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:01.185375 master-0 kubenswrapper[24928]: I1205 10:54:01.185316 24928 status_manager.go:851] "Failed to get status for pod" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" pod="openshift-console/console-74f96dcf4d-9gskd" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-console/pods/console-74f96dcf4d-9gskd\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:01.558199 master-0 kubenswrapper[24928]: E1205 10:54:01.558119 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="1.6s" Dec 05 10:54:02.189160 master-0 kubenswrapper[24928]: I1205 10:54:02.189086 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:02.189160 master-0 kubenswrapper[24928]: I1205 10:54:02.189144 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:02.190049 master-0 kubenswrapper[24928]: E1205 10:54:02.189967 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:54:02.941188 master-0 kubenswrapper[24928]: E1205 10:54:02.941104 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:54:02Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:54:02Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:54:02Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T10:54:02Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"master-0\": Patch \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0/status?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:02.941941 master-0 kubenswrapper[24928]: E1205 10:54:02.941638 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:02.942295 master-0 kubenswrapper[24928]: E1205 10:54:02.942246 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:02.942724 master-0 kubenswrapper[24928]: E1205 10:54:02.942684 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:02.943346 master-0 kubenswrapper[24928]: E1205 10:54:02.943301 24928 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"master-0\": Get \"https://api-int.sno.openstack.lab:6443/api/v1/nodes/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:02.943346 master-0 kubenswrapper[24928]: E1205 10:54:02.943327 24928 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 10:54:03.159661 master-0 kubenswrapper[24928]: E1205 10:54:03.159587 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="3.2s" Dec 05 10:54:04.904390 master-0 kubenswrapper[24928]: I1205 10:54:04.904280 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:54:04.905239 master-0 kubenswrapper[24928]: I1205 10:54:04.904399 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:54:05.203964 master-0 kubenswrapper[24928]: I1205 10:54:05.203851 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:54:05.206054 master-0 kubenswrapper[24928]: I1205 10:54:05.205953 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:05.207211 master-0 kubenswrapper[24928]: I1205 10:54:05.207129 24928 status_manager.go:851] "Failed to get status for pod" podUID="f8788a6e4f698a66d5d8f663c8089807" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:05.208077 master-0 kubenswrapper[24928]: I1205 10:54:05.208011 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:05.209246 master-0 kubenswrapper[24928]: I1205 10:54:05.209151 24928 status_manager.go:851] "Failed to get status for pod" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" pod="openshift-console/console-74f96dcf4d-9gskd" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-console/pods/console-74f96dcf4d-9gskd\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:05.224469 master-0 kubenswrapper[24928]: I1205 10:54:05.224349 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:05.224469 master-0 kubenswrapper[24928]: I1205 10:54:05.224404 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:05.226087 master-0 kubenswrapper[24928]: E1205 10:54:05.226000 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:54:05.226806 master-0 kubenswrapper[24928]: I1205 10:54:05.226763 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:54:05.252657 master-0 kubenswrapper[24928]: W1205 10:54:05.252581 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda369cadf0161d66f2936cdea3ded59b7.slice/crio-5d5c5d096c7ad6cde9f6bcb86ca062e81f05e3dce9a4d346bf730461efca0d4f WatchSource:0}: Error finding container 5d5c5d096c7ad6cde9f6bcb86ca062e81f05e3dce9a4d346bf730461efca0d4f: Status 404 returned error can't find the container with id 5d5c5d096c7ad6cde9f6bcb86ca062e81f05e3dce9a4d346bf730461efca0d4f Dec 05 10:54:06.221511 master-0 kubenswrapper[24928]: I1205 10:54:06.221374 24928 generic.go:334] "Generic (PLEG): container finished" podID="a369cadf0161d66f2936cdea3ded59b7" containerID="5132d86e4e2bafc3b4c03c5d3a7f70d47675331e8a44adc6bccddf638855598e" exitCode=0 Dec 05 10:54:06.221954 master-0 kubenswrapper[24928]: I1205 10:54:06.221475 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"a369cadf0161d66f2936cdea3ded59b7","Type":"ContainerDied","Data":"5132d86e4e2bafc3b4c03c5d3a7f70d47675331e8a44adc6bccddf638855598e"} Dec 05 10:54:06.221954 master-0 kubenswrapper[24928]: I1205 10:54:06.221570 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"a369cadf0161d66f2936cdea3ded59b7","Type":"ContainerStarted","Data":"5d5c5d096c7ad6cde9f6bcb86ca062e81f05e3dce9a4d346bf730461efca0d4f"} Dec 05 10:54:06.222041 master-0 kubenswrapper[24928]: I1205 10:54:06.222008 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:06.222084 master-0 kubenswrapper[24928]: I1205 10:54:06.222047 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:06.222901 master-0 kubenswrapper[24928]: E1205 10:54:06.222864 24928 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:54:06.222947 master-0 kubenswrapper[24928]: I1205 10:54:06.222872 24928 status_manager.go:851] "Failed to get status for pod" podUID="b83ccd6fa217a93a2c607d0109896ef8" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:06.223918 master-0 kubenswrapper[24928]: I1205 10:54:06.223592 24928 status_manager.go:851] "Failed to get status for pod" podUID="f8788a6e4f698a66d5d8f663c8089807" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:06.224665 master-0 kubenswrapper[24928]: I1205 10:54:06.224602 24928 status_manager.go:851] "Failed to get status for pod" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" pod="openshift-kube-apiserver/installer-5-master-0" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-5-master-0\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:06.225532 master-0 kubenswrapper[24928]: I1205 10:54:06.225486 24928 status_manager.go:851] "Failed to get status for pod" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" pod="openshift-console/console-74f96dcf4d-9gskd" err="Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-console/pods/console-74f96dcf4d-9gskd\": dial tcp 192.168.32.10:6443: connect: connection refused" Dec 05 10:54:06.361598 master-0 kubenswrapper[24928]: E1205 10:54:06.361521 24928 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/master-0?timeout=10s\": dial tcp 192.168.32.10:6443: connect: connection refused" interval="6.4s" Dec 05 10:54:07.243342 master-0 kubenswrapper[24928]: I1205 10:54:07.243092 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"a369cadf0161d66f2936cdea3ded59b7","Type":"ContainerStarted","Data":"e6eb4591ccbc52e4e29d43a03a4a12d9503d1518cbfdd97974e2723c416ce572"} Dec 05 10:54:07.243342 master-0 kubenswrapper[24928]: I1205 10:54:07.243148 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"a369cadf0161d66f2936cdea3ded59b7","Type":"ContainerStarted","Data":"f3bd46c4d03f23856889f42f059cba33bdae0b85f9b80b702530d8c0a7671c55"} Dec 05 10:54:07.243342 master-0 kubenswrapper[24928]: I1205 10:54:07.243162 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"a369cadf0161d66f2936cdea3ded59b7","Type":"ContainerStarted","Data":"a01d650adfe1f5d8339d11eebb1d1cbfd897a778c84ec9ef0d4e91ee0faeca08"} Dec 05 10:54:08.251792 master-0 kubenswrapper[24928]: I1205 10:54:08.251732 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"a369cadf0161d66f2936cdea3ded59b7","Type":"ContainerStarted","Data":"36446f9fbd9cc9da1d8d5a7812037c83f93cb6a7ba4d88d11b9bcb4699f79ffb"} Dec 05 10:54:08.251792 master-0 kubenswrapper[24928]: I1205 10:54:08.251781 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" event={"ID":"a369cadf0161d66f2936cdea3ded59b7","Type":"ContainerStarted","Data":"d0482e0fad95d6016af5598b22c91f9d69295790989a933990132631050c3542"} Dec 05 10:54:08.252448 master-0 kubenswrapper[24928]: I1205 10:54:08.252015 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:08.252448 master-0 kubenswrapper[24928]: I1205 10:54:08.252031 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:08.252448 master-0 kubenswrapper[24928]: I1205 10:54:08.252281 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:54:09.310342 master-0 kubenswrapper[24928]: I1205 10:54:09.310269 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:54:09.310342 master-0 kubenswrapper[24928]: I1205 10:54:09.310330 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:54:09.310342 master-0 kubenswrapper[24928]: I1205 10:54:09.310342 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:54:09.310996 master-0 kubenswrapper[24928]: I1205 10:54:09.310629 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:54:09.310996 master-0 kubenswrapper[24928]: I1205 10:54:09.310685 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:09.310996 master-0 kubenswrapper[24928]: I1205 10:54:09.310703 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:09.314068 master-0 kubenswrapper[24928]: I1205 10:54:09.313995 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:54:10.227553 master-0 kubenswrapper[24928]: I1205 10:54:10.227448 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:54:10.227553 master-0 kubenswrapper[24928]: I1205 10:54:10.227547 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:54:10.233798 master-0 kubenswrapper[24928]: I1205 10:54:10.233722 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:54:12.310454 master-0 kubenswrapper[24928]: I1205 10:54:12.310377 24928 patch_prober.go:28] interesting pod/kube-controller-manager-master-0 container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 10:54:12.311107 master-0 kubenswrapper[24928]: I1205 10:54:12.310469 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="f8788a6e4f698a66d5d8f663c8089807" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 10:54:13.264930 master-0 kubenswrapper[24928]: I1205 10:54:13.264838 24928 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:54:13.286236 master-0 kubenswrapper[24928]: I1205 10:54:13.286158 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:13.286236 master-0 kubenswrapper[24928]: I1205 10:54:13.286199 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:13.290244 master-0 kubenswrapper[24928]: I1205 10:54:13.290195 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:54:13.388451 master-0 kubenswrapper[24928]: I1205 10:54:13.387475 24928 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-master-0" oldPodUID="a369cadf0161d66f2936cdea3ded59b7" podUID="f9d62fb8-e9e3-46c1-a7e1-a3502ff21c9a" Dec 05 10:54:14.296032 master-0 kubenswrapper[24928]: I1205 10:54:14.295925 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:14.296032 master-0 kubenswrapper[24928]: I1205 10:54:14.295981 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:14.298750 master-0 kubenswrapper[24928]: I1205 10:54:14.298687 24928 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-master-0" oldPodUID="a369cadf0161d66f2936cdea3ded59b7" podUID="f9d62fb8-e9e3-46c1-a7e1-a3502ff21c9a" Dec 05 10:54:14.325802 master-0 kubenswrapper[24928]: I1205 10:54:14.325724 24928 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:54:14.903962 master-0 kubenswrapper[24928]: I1205 10:54:14.903880 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:54:14.903962 master-0 kubenswrapper[24928]: I1205 10:54:14.903957 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:54:15.303223 master-0 kubenswrapper[24928]: I1205 10:54:15.303117 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:15.303223 master-0 kubenswrapper[24928]: I1205 10:54:15.303160 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:15.309583 master-0 kubenswrapper[24928]: I1205 10:54:15.309515 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:54:15.312675 master-0 kubenswrapper[24928]: I1205 10:54:15.312587 24928 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" oldPodUID="f8788a6e4f698a66d5d8f663c8089807" podUID="d4b830cb-6904-4e50-bdf7-c898f94ceb4f" Dec 05 10:54:16.309168 master-0 kubenswrapper[24928]: I1205 10:54:16.309082 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:16.309168 master-0 kubenswrapper[24928]: I1205 10:54:16.309126 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:19.319233 master-0 kubenswrapper[24928]: I1205 10:54:19.319023 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:54:19.320513 master-0 kubenswrapper[24928]: I1205 10:54:19.319521 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:19.320513 master-0 kubenswrapper[24928]: I1205 10:54:19.319547 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:19.326888 master-0 kubenswrapper[24928]: I1205 10:54:19.326802 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" Dec 05 10:54:19.332851 master-0 kubenswrapper[24928]: I1205 10:54:19.332784 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:19.332851 master-0 kubenswrapper[24928]: I1205 10:54:19.332829 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:20.213440 master-0 kubenswrapper[24928]: I1205 10:54:20.213309 24928 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" oldPodUID="f8788a6e4f698a66d5d8f663c8089807" podUID="d4b830cb-6904-4e50-bdf7-c898f94ceb4f" Dec 05 10:54:22.844535 master-0 kubenswrapper[24928]: I1205 10:54:22.844440 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 10:54:23.064155 master-0 kubenswrapper[24928]: I1205 10:54:23.064053 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 10:54:23.105328 master-0 kubenswrapper[24928]: I1205 10:54:23.105151 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client" Dec 05 10:54:23.255524 master-0 kubenswrapper[24928]: I1205 10:54:23.255460 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Dec 05 10:54:24.099648 master-0 kubenswrapper[24928]: I1205 10:54:24.099512 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 10:54:24.153223 master-0 kubenswrapper[24928]: I1205 10:54:24.153118 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Dec 05 10:54:24.603764 master-0 kubenswrapper[24928]: I1205 10:54:24.603667 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"baremetal-kube-rbac-proxy" Dec 05 10:54:24.717875 master-0 kubenswrapper[24928]: I1205 10:54:24.717763 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 10:54:24.801235 master-0 kubenswrapper[24928]: I1205 10:54:24.801164 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 10:54:24.904214 master-0 kubenswrapper[24928]: I1205 10:54:24.903989 24928 patch_prober.go:28] interesting pod/console-79cdddb8b4-mwjwx container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" start-of-body= Dec 05 10:54:24.904214 master-0 kubenswrapper[24928]: I1205 10:54:24.904068 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" probeResult="failure" output="Get \"https://10.128.0.87:8443/health\": dial tcp 10.128.0.87:8443: connect: connection refused" Dec 05 10:54:24.905790 master-0 kubenswrapper[24928]: I1205 10:54:24.905714 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-l9vpw" Dec 05 10:54:25.082470 master-0 kubenswrapper[24928]: I1205 10:54:25.082361 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 10:54:25.145081 master-0 kubenswrapper[24928]: I1205 10:54:25.144978 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 10:54:25.335822 master-0 kubenswrapper[24928]: I1205 10:54:25.335681 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 10:54:25.446035 master-0 kubenswrapper[24928]: I1205 10:54:25.445786 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 10:54:25.901235 master-0 kubenswrapper[24928]: I1205 10:54:25.901168 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-node-tuning-operator"/"node-tuning-operator-tls" Dec 05 10:54:25.972099 master-0 kubenswrapper[24928]: I1205 10:54:25.972037 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-6j2qd" Dec 05 10:54:26.025585 master-0 kubenswrapper[24928]: I1205 10:54:26.025522 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Dec 05 10:54:26.056764 master-0 kubenswrapper[24928]: I1205 10:54:26.056716 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 10:54:26.175272 master-0 kubenswrapper[24928]: I1205 10:54:26.175135 24928 scope.go:117] "RemoveContainer" containerID="edac655814a82eb2075cd815c3a2e8122778082bf81de28d1d2bd7119b304f24" Dec 05 10:54:26.196178 master-0 kubenswrapper[24928]: I1205 10:54:26.196132 24928 scope.go:117] "RemoveContainer" containerID="47bef9607325414e871a594b06d4250139a33f93aa8b2368ce9838e31d093051" Dec 05 10:54:26.268360 master-0 kubenswrapper[24928]: I1205 10:54:26.268307 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 10:54:26.281353 master-0 kubenswrapper[24928]: I1205 10:54:26.281295 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 10:54:26.295540 master-0 kubenswrapper[24928]: I1205 10:54:26.295505 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Dec 05 10:54:26.323409 master-0 kubenswrapper[24928]: I1205 10:54:26.323350 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-storage-operator"/"cluster-storage-operator-dockercfg-g78vb" Dec 05 10:54:26.339059 master-0 kubenswrapper[24928]: I1205 10:54:26.338952 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 10:54:26.376646 master-0 kubenswrapper[24928]: I1205 10:54:26.376568 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 10:54:26.408193 master-0 kubenswrapper[24928]: I1205 10:54:26.408125 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 10:54:26.423769 master-0 kubenswrapper[24928]: I1205 10:54:26.423725 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 10:54:26.545538 master-0 kubenswrapper[24928]: I1205 10:54:26.545467 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 10:54:26.614010 master-0 kubenswrapper[24928]: I1205 10:54:26.613932 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 10:54:26.640676 master-0 kubenswrapper[24928]: I1205 10:54:26.640576 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 10:54:26.737278 master-0 kubenswrapper[24928]: I1205 10:54:26.736917 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-qkdkt" Dec 05 10:54:26.983247 master-0 kubenswrapper[24928]: I1205 10:54:26.983095 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 10:54:27.011527 master-0 kubenswrapper[24928]: I1205 10:54:27.011202 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 10:54:27.097810 master-0 kubenswrapper[24928]: I1205 10:54:27.097753 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 10:54:27.170815 master-0 kubenswrapper[24928]: I1205 10:54:27.170743 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"trusted-ca-bundle" Dec 05 10:54:27.298566 master-0 kubenswrapper[24928]: I1205 10:54:27.298472 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"federate-client-certs" Dec 05 10:54:27.299195 master-0 kubenswrapper[24928]: I1205 10:54:27.298826 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 10:54:27.303513 master-0 kubenswrapper[24928]: I1205 10:54:27.303469 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client-dockercfg-2kvvs" Dec 05 10:54:27.366195 master-0 kubenswrapper[24928]: I1205 10:54:27.366103 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 10:54:27.395235 master-0 kubenswrapper[24928]: I1205 10:54:27.395156 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 10:54:27.398979 master-0 kubenswrapper[24928]: I1205 10:54:27.398922 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 10:54:27.418538 master-0 kubenswrapper[24928]: I1205 10:54:27.418481 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 10:54:27.551352 master-0 kubenswrapper[24928]: I1205 10:54:27.551231 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-baremetal-webhook-server-cert" Dec 05 10:54:27.595480 master-0 kubenswrapper[24928]: I1205 10:54:27.595410 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-insights"/"openshift-insights-serving-cert" Dec 05 10:54:27.635252 master-0 kubenswrapper[24928]: I1205 10:54:27.635199 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 10:54:27.694169 master-0 kubenswrapper[24928]: I1205 10:54:27.694106 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 10:54:27.729304 master-0 kubenswrapper[24928]: I1205 10:54:27.729215 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 10:54:27.905938 master-0 kubenswrapper[24928]: I1205 10:54:27.905735 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 10:54:27.923210 master-0 kubenswrapper[24928]: I1205 10:54:27.923155 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-storage-operator"/"cluster-storage-operator-serving-cert" Dec 05 10:54:28.045098 master-0 kubenswrapper[24928]: I1205 10:54:28.045027 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 10:54:28.051011 master-0 kubenswrapper[24928]: I1205 10:54:28.050957 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 10:54:28.562786 master-0 kubenswrapper[24928]: I1205 10:54:28.562710 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 10:54:28.567558 master-0 kubenswrapper[24928]: I1205 10:54:28.567515 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 10:54:28.604288 master-0 kubenswrapper[24928]: I1205 10:54:28.604234 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client-kube-rbac-proxy-config" Dec 05 10:54:28.604525 master-0 kubenswrapper[24928]: I1205 10:54:28.604318 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 10:54:28.636506 master-0 kubenswrapper[24928]: I1205 10:54:28.636443 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-2bqcq" Dec 05 10:54:28.677912 master-0 kubenswrapper[24928]: I1205 10:54:28.677850 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:54:28.690678 master-0 kubenswrapper[24928]: I1205 10:54:28.690607 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 10:54:28.699392 master-0 kubenswrapper[24928]: I1205 10:54:28.699347 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-s9q7m" Dec 05 10:54:28.728095 master-0 kubenswrapper[24928]: I1205 10:54:28.728044 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-olm-operator"/"kube-root-ca.crt" Dec 05 10:54:28.777296 master-0 kubenswrapper[24928]: I1205 10:54:28.777239 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 10:54:28.802126 master-0 kubenswrapper[24928]: I1205 10:54:28.802067 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-storage-operator"/"openshift-service-ca.crt" Dec 05 10:54:28.867609 master-0 kubenswrapper[24928]: I1205 10:54:28.867560 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 10:54:29.107950 master-0 kubenswrapper[24928]: I1205 10:54:29.107837 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 10:54:29.111311 master-0 kubenswrapper[24928]: I1205 10:54:29.111287 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-m8wb4" Dec 05 10:54:29.131565 master-0 kubenswrapper[24928]: I1205 10:54:29.131506 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-credential-operator"/"cco-trusted-ca" Dec 05 10:54:29.191951 master-0 kubenswrapper[24928]: I1205 10:54:29.191896 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"telemeter-client-tls" Dec 05 10:54:29.259447 master-0 kubenswrapper[24928]: I1205 10:54:29.259375 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Dec 05 10:54:29.277078 master-0 kubenswrapper[24928]: I1205 10:54:29.277039 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 10:54:29.315203 master-0 kubenswrapper[24928]: I1205 10:54:29.315163 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 10:54:29.384075 master-0 kubenswrapper[24928]: I1205 10:54:29.383940 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 10:54:29.386282 master-0 kubenswrapper[24928]: I1205 10:54:29.386199 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 10:54:29.395122 master-0 kubenswrapper[24928]: I1205 10:54:29.395071 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Dec 05 10:54:29.411892 master-0 kubenswrapper[24928]: I1205 10:54:29.411851 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 10:54:29.431044 master-0 kubenswrapper[24928]: I1205 10:54:29.430992 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"kube-rbac-proxy" Dec 05 10:54:29.623984 master-0 kubenswrapper[24928]: I1205 10:54:29.623895 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 10:54:29.914893 master-0 kubenswrapper[24928]: I1205 10:54:29.914789 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 10:54:29.950241 master-0 kubenswrapper[24928]: I1205 10:54:29.950165 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 10:54:29.993241 master-0 kubenswrapper[24928]: I1205 10:54:29.993168 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 10:54:30.041189 master-0 kubenswrapper[24928]: I1205 10:54:30.041132 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 10:54:30.081437 master-0 kubenswrapper[24928]: I1205 10:54:30.081376 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 10:54:30.094731 master-0 kubenswrapper[24928]: I1205 10:54:30.094668 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 10:54:30.182847 master-0 kubenswrapper[24928]: I1205 10:54:30.182656 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-autoscaler-operator-cert" Dec 05 10:54:30.218766 master-0 kubenswrapper[24928]: I1205 10:54:30.218705 24928 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 10:54:30.276377 master-0 kubenswrapper[24928]: I1205 10:54:30.276258 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-credential-operator"/"cloud-credential-operator-serving-cert" Dec 05 10:54:30.304487 master-0 kubenswrapper[24928]: I1205 10:54:30.304410 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 10:54:30.347055 master-0 kubenswrapper[24928]: I1205 10:54:30.346973 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemeter-trusted-ca-bundle-56c9b9fa8d9gs" Dec 05 10:54:30.351694 master-0 kubenswrapper[24928]: I1205 10:54:30.351631 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy-cluster-autoscaler-operator" Dec 05 10:54:30.540630 master-0 kubenswrapper[24928]: I1205 10:54:30.540555 24928 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 10:54:30.577099 master-0 kubenswrapper[24928]: I1205 10:54:30.577014 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 10:54:30.586793 master-0 kubenswrapper[24928]: I1205 10:54:30.586710 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-2jmqj" Dec 05 10:54:30.603772 master-0 kubenswrapper[24928]: I1205 10:54:30.603695 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-catalogd"/"catalogd-trusted-ca-bundle" Dec 05 10:54:30.776928 master-0 kubenswrapper[24928]: I1205 10:54:30.776804 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 10:54:30.819108 master-0 kubenswrapper[24928]: I1205 10:54:30.818965 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:54:30.860376 master-0 kubenswrapper[24928]: I1205 10:54:30.860034 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 10:54:30.881029 master-0 kubenswrapper[24928]: I1205 10:54:30.880930 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 10:54:31.005341 master-0 kubenswrapper[24928]: I1205 10:54:31.005256 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 10:54:31.032895 master-0 kubenswrapper[24928]: I1205 10:54:31.032813 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"whereabouts-config" Dec 05 10:54:31.041828 master-0 kubenswrapper[24928]: I1205 10:54:31.041767 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 10:54:31.145283 master-0 kubenswrapper[24928]: I1205 10:54:31.145096 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 10:54:31.182917 master-0 kubenswrapper[24928]: I1205 10:54:31.182816 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 10:54:31.256961 master-0 kubenswrapper[24928]: I1205 10:54:31.256879 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 10:54:31.305087 master-0 kubenswrapper[24928]: I1205 10:54:31.305040 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 10:54:31.460256 master-0 kubenswrapper[24928]: I1205 10:54:31.460124 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-catalogd"/"catalogserver-cert" Dec 05 10:54:31.471306 master-0 kubenswrapper[24928]: I1205 10:54:31.471253 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 10:54:31.564654 master-0 kubenswrapper[24928]: I1205 10:54:31.564597 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 10:54:31.580363 master-0 kubenswrapper[24928]: I1205 10:54:31.580233 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-brtjg" Dec 05 10:54:31.588160 master-0 kubenswrapper[24928]: I1205 10:54:31.588105 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 10:54:31.617135 master-0 kubenswrapper[24928]: I1205 10:54:31.617073 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 10:54:31.684316 master-0 kubenswrapper[24928]: I1205 10:54:31.684262 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 10:54:31.725437 master-0 kubenswrapper[24928]: I1205 10:54:31.725220 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-storage-operator"/"kube-root-ca.crt" Dec 05 10:54:31.784444 master-0 kubenswrapper[24928]: I1205 10:54:31.784368 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-fvzgd" Dec 05 10:54:31.788832 master-0 kubenswrapper[24928]: I1205 10:54:31.788768 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 10:54:31.793815 master-0 kubenswrapper[24928]: I1205 10:54:31.793772 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 10:54:31.810793 master-0 kubenswrapper[24928]: I1205 10:54:31.810729 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 10:54:31.811057 master-0 kubenswrapper[24928]: I1205 10:54:31.811011 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-controller"/"kube-root-ca.crt" Dec 05 10:54:31.818847 master-0 kubenswrapper[24928]: I1205 10:54:31.818803 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 10:54:31.897075 master-0 kubenswrapper[24928]: I1205 10:54:31.897001 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 10:54:31.940286 master-0 kubenswrapper[24928]: I1205 10:54:31.940225 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-zlzd6" Dec 05 10:54:32.071020 master-0 kubenswrapper[24928]: I1205 10:54:32.070946 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 10:54:32.139918 master-0 kubenswrapper[24928]: I1205 10:54:32.139852 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-2qdv2" Dec 05 10:54:32.166713 master-0 kubenswrapper[24928]: I1205 10:54:32.166632 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-baremetal-operator-dockercfg-h7kbf" Dec 05 10:54:32.171038 master-0 kubenswrapper[24928]: I1205 10:54:32.170980 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 10:54:32.238284 master-0 kubenswrapper[24928]: I1205 10:54:32.238204 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Dec 05 10:54:32.273938 master-0 kubenswrapper[24928]: I1205 10:54:32.273873 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 10:54:32.338029 master-0 kubenswrapper[24928]: I1205 10:54:32.337894 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 10:54:32.340103 master-0 kubenswrapper[24928]: I1205 10:54:32.340052 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-controller"/"openshift-service-ca.crt" Dec 05 10:54:32.342924 master-0 kubenswrapper[24928]: I1205 10:54:32.342887 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 10:54:32.398687 master-0 kubenswrapper[24928]: I1205 10:54:32.398618 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 10:54:32.540472 master-0 kubenswrapper[24928]: I1205 10:54:32.540390 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 10:54:32.591070 master-0 kubenswrapper[24928]: I1205 10:54:32.590867 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 10:54:32.620181 master-0 kubenswrapper[24928]: I1205 10:54:32.620121 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Dec 05 10:54:32.823372 master-0 kubenswrapper[24928]: I1205 10:54:32.823300 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-5ll0c5ruaqfm2" Dec 05 10:54:32.860471 master-0 kubenswrapper[24928]: I1205 10:54:32.860291 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Dec 05 10:54:32.947865 master-0 kubenswrapper[24928]: I1205 10:54:32.947783 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 10:54:33.010939 master-0 kubenswrapper[24928]: I1205 10:54:33.010872 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 10:54:33.023324 master-0 kubenswrapper[24928]: I1205 10:54:33.023271 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 10:54:33.057541 master-0 kubenswrapper[24928]: I1205 10:54:33.057486 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 10:54:33.088415 master-0 kubenswrapper[24928]: I1205 10:54:33.088349 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Dec 05 10:54:33.103584 master-0 kubenswrapper[24928]: I1205 10:54:33.103518 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 10:54:33.107791 master-0 kubenswrapper[24928]: I1205 10:54:33.106262 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 10:54:33.195737 master-0 kubenswrapper[24928]: I1205 10:54:33.195609 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-catalogd"/"openshift-service-ca.crt" Dec 05 10:54:33.282599 master-0 kubenswrapper[24928]: I1205 10:54:33.282521 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 10:54:33.305657 master-0 kubenswrapper[24928]: I1205 10:54:33.305603 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 10:54:33.346780 master-0 kubenswrapper[24928]: I1205 10:54:33.346709 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-credential-operator"/"kube-root-ca.crt" Dec 05 10:54:33.370452 master-0 kubenswrapper[24928]: I1205 10:54:33.370353 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 10:54:33.466483 master-0 kubenswrapper[24928]: I1205 10:54:33.466323 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 10:54:33.466750 master-0 kubenswrapper[24928]: I1205 10:54:33.466702 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-insights"/"operator-dockercfg-lngmk" Dec 05 10:54:33.528452 master-0 kubenswrapper[24928]: I1205 10:54:33.528383 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 10:54:33.609923 master-0 kubenswrapper[24928]: I1205 10:54:33.609837 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 10:54:33.650565 master-0 kubenswrapper[24928]: I1205 10:54:33.650508 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 10:54:33.732696 master-0 kubenswrapper[24928]: I1205 10:54:33.732570 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-credential-operator"/"openshift-service-ca.crt" Dec 05 10:54:33.751366 master-0 kubenswrapper[24928]: I1205 10:54:33.751309 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-9zg4x" Dec 05 10:54:33.775884 master-0 kubenswrapper[24928]: I1205 10:54:33.775835 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 10:54:33.820376 master-0 kubenswrapper[24928]: I1205 10:54:33.820291 24928 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 10:54:33.903993 master-0 kubenswrapper[24928]: I1205 10:54:33.903912 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-controller"/"operator-controller-trusted-ca-bundle" Dec 05 10:54:33.936673 master-0 kubenswrapper[24928]: I1205 10:54:33.936605 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-x98dp" Dec 05 10:54:34.099309 master-0 kubenswrapper[24928]: I1205 10:54:34.099240 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 10:54:34.143043 master-0 kubenswrapper[24928]: I1205 10:54:34.142958 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 10:54:34.167662 master-0 kubenswrapper[24928]: I1205 10:54:34.167583 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 10:54:34.226787 master-0 kubenswrapper[24928]: I1205 10:54:34.226722 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 10:54:34.285375 master-0 kubenswrapper[24928]: I1205 10:54:34.285309 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"cluster-baremetal-operator-images" Dec 05 10:54:34.325600 master-0 kubenswrapper[24928]: I1205 10:54:34.325529 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 10:54:34.359732 master-0 kubenswrapper[24928]: I1205 10:54:34.359584 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-autoscaler-operator-dockercfg-vwmqb" Dec 05 10:54:34.361618 master-0 kubenswrapper[24928]: I1205 10:54:34.361584 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-vp7fw" Dec 05 10:54:34.454782 master-0 kubenswrapper[24928]: I1205 10:54:34.454700 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-pz2dt" Dec 05 10:54:34.460106 master-0 kubenswrapper[24928]: I1205 10:54:34.460074 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 10:54:34.472345 master-0 kubenswrapper[24928]: I1205 10:54:34.472301 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 10:54:34.665857 master-0 kubenswrapper[24928]: I1205 10:54:34.665674 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 10:54:34.741100 master-0 kubenswrapper[24928]: I1205 10:54:34.741036 24928 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 10:54:34.743233 master-0 kubenswrapper[24928]: I1205 10:54:34.743167 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" podStartSLOduration=42.743150034 podStartE2EDuration="42.743150034s" podCreationTimestamp="2025-12-05 10:53:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:54:13.274006184 +0000 UTC m=+413.277200035" watchObservedRunningTime="2025-12-05 10:54:34.743150034 +0000 UTC m=+434.746343905" Dec 05 10:54:34.750349 master-0 kubenswrapper[24928]: I1205 10:54:34.750052 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-master-0","openshift-console/console-74f96dcf4d-9gskd","openshift-kube-apiserver/kube-apiserver-master-0"] Dec 05 10:54:34.750541 master-0 kubenswrapper[24928]: I1205 10:54:34.750521 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-52wjg","openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv","openshift-marketplace/community-operators-mcjzc","openshift-marketplace/redhat-operators-8pb58","openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq","openshift-marketplace/redhat-marketplace-l4grl","openshift-monitoring/prometheus-k8s-0","openshift-kube-apiserver/kube-apiserver-master-0","openshift-controller-manager/controller-manager-8f9584d48-fblwk","openshift-console/console-74977ddd8b-dkrkh","openshift-monitoring/thanos-querier-598896584f-9pd95","openshift-monitoring/alertmanager-main-0","openshift-monitoring/metrics-server-64494f74c5-sqgmf","openshift-kube-controller-manager/kube-controller-manager-master-0"] Dec 05 10:54:34.750864 master-0 kubenswrapper[24928]: E1205 10:54:34.750825 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" Dec 05 10:54:34.750923 master-0 kubenswrapper[24928]: I1205 10:54:34.750865 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" Dec 05 10:54:34.750923 master-0 kubenswrapper[24928]: E1205 10:54:34.750899 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" containerName="installer" Dec 05 10:54:34.750923 master-0 kubenswrapper[24928]: I1205 10:54:34.750913 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" containerName="installer" Dec 05 10:54:34.751062 master-0 kubenswrapper[24928]: I1205 10:54:34.750917 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:34.751062 master-0 kubenswrapper[24928]: I1205 10:54:34.750948 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="172dd88f-0e78-4edf-8ed7-0112a64775ca" Dec 05 10:54:34.751062 master-0 kubenswrapper[24928]: I1205 10:54:34.750985 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:34.751062 master-0 kubenswrapper[24928]: I1205 10:54:34.751022 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-master-0" podUID="16bc8db9-9fd8-4044-885a-910a37445387" Dec 05 10:54:34.751232 master-0 kubenswrapper[24928]: I1205 10:54:34.751184 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" Dec 05 10:54:34.751232 master-0 kubenswrapper[24928]: I1205 10:54:34.751217 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" Dec 05 10:54:34.751325 master-0 kubenswrapper[24928]: I1205 10:54:34.751259 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d237cdd-9aa8-4a12-b244-1dc1b4c456d0" containerName="installer" Dec 05 10:54:34.751533 master-0 kubenswrapper[24928]: E1205 10:54:34.751509 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" Dec 05 10:54:34.751533 master-0 kubenswrapper[24928]: I1205 10:54:34.751531 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" containerName="console" Dec 05 10:54:34.754334 master-0 kubenswrapper[24928]: I1205 10:54:34.753487 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:34.755672 master-0 kubenswrapper[24928]: I1205 10:54:34.755638 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:34.757811 master-0 kubenswrapper[24928]: I1205 10:54:34.757770 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:34.759860 master-0 kubenswrapper[24928]: I1205 10:54:34.759822 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:34.763274 master-0 kubenswrapper[24928]: I1205 10:54:34.763234 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.765629 master-0 kubenswrapper[24928]: I1205 10:54:34.765551 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.765805 master-0 kubenswrapper[24928]: I1205 10:54:34.765726 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.767546 master-0 kubenswrapper[24928]: I1205 10:54:34.767496 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.768149 master-0 kubenswrapper[24928]: I1205 10:54:34.768088 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Dec 05 10:54:34.768279 master-0 kubenswrapper[24928]: I1205 10:54:34.768139 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Dec 05 10:54:34.768279 master-0 kubenswrapper[24928]: I1205 10:54:34.768101 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Dec 05 10:54:34.768279 master-0 kubenswrapper[24928]: I1205 10:54:34.768238 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Dec 05 10:54:34.768536 master-0 kubenswrapper[24928]: I1205 10:54:34.768246 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Dec 05 10:54:34.768715 master-0 kubenswrapper[24928]: I1205 10:54:34.768652 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 10:54:34.768821 master-0 kubenswrapper[24928]: I1205 10:54:34.768807 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Dec 05 10:54:34.769566 master-0 kubenswrapper[24928]: I1205 10:54:34.769508 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Dec 05 10:54:34.770243 master-0 kubenswrapper[24928]: I1205 10:54:34.769823 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 10:54:34.770243 master-0 kubenswrapper[24928]: I1205 10:54:34.769829 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 10:54:34.770243 master-0 kubenswrapper[24928]: I1205 10:54:34.770033 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-4d9vt0h39vbq9" Dec 05 10:54:34.770243 master-0 kubenswrapper[24928]: I1205 10:54:34.770159 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Dec 05 10:54:34.771366 master-0 kubenswrapper[24928]: I1205 10:54:34.771117 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 10:54:34.771366 master-0 kubenswrapper[24928]: I1205 10:54:34.771237 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-44p5j" Dec 05 10:54:34.771598 master-0 kubenswrapper[24928]: I1205 10:54:34.771558 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 10:54:34.772119 master-0 kubenswrapper[24928]: I1205 10:54:34.771715 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 10:54:34.772119 master-0 kubenswrapper[24928]: I1205 10:54:34.771947 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-4zslp" Dec 05 10:54:34.773243 master-0 kubenswrapper[24928]: I1205 10:54:34.772838 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 10:54:34.773243 master-0 kubenswrapper[24928]: I1205 10:54:34.772885 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 10:54:34.773474 master-0 kubenswrapper[24928]: I1205 10:54:34.773265 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Dec 05 10:54:34.773566 master-0 kubenswrapper[24928]: I1205 10:54:34.773505 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 10:54:34.773637 master-0 kubenswrapper[24928]: I1205 10:54:34.773600 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 10:54:34.773769 master-0 kubenswrapper[24928]: I1205 10:54:34.773739 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 10:54:34.773869 master-0 kubenswrapper[24928]: I1205 10:54:34.773847 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 10:54:34.773952 master-0 kubenswrapper[24928]: I1205 10:54:34.773849 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 10:54:34.773952 master-0 kubenswrapper[24928]: I1205 10:54:34.773881 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 10:54:34.773952 master-0 kubenswrapper[24928]: I1205 10:54:34.773898 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 10:54:34.774157 master-0 kubenswrapper[24928]: I1205 10:54:34.773967 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 10:54:34.775679 master-0 kubenswrapper[24928]: I1205 10:54:34.774230 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 10:54:34.775679 master-0 kubenswrapper[24928]: I1205 10:54:34.774249 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 10:54:34.775679 master-0 kubenswrapper[24928]: I1205 10:54:34.775217 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 10:54:34.776152 master-0 kubenswrapper[24928]: I1205 10:54:34.776105 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 10:54:34.776241 master-0 kubenswrapper[24928]: I1205 10:54:34.776052 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 10:54:34.778255 master-0 kubenswrapper[24928]: I1205 10:54:34.778195 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.788271 master-0 kubenswrapper[24928]: I1205 10:54:34.788174 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 10:54:34.791726 master-0 kubenswrapper[24928]: I1205 10:54:34.791678 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.792004 master-0 kubenswrapper[24928]: I1205 10:54:34.791926 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Dec 05 10:54:34.795263 master-0 kubenswrapper[24928]: I1205 10:54:34.795222 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 10:54:34.797348 master-0 kubenswrapper[24928]: I1205 10:54:34.797283 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Dec 05 10:54:34.797854 master-0 kubenswrapper[24928]: I1205 10:54:34.797802 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.798074 master-0 kubenswrapper[24928]: I1205 10:54:34.798031 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-79cdddb8b4-mwjwx"] Dec 05 10:54:34.798131 master-0 kubenswrapper[24928]: I1205 10:54:34.798106 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-master-0" Dec 05 10:54:34.798351 master-0 kubenswrapper[24928]: I1205 10:54:34.798317 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.798861 master-0 kubenswrapper[24928]: I1205 10:54:34.798833 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Dec 05 10:54:34.799392 master-0 kubenswrapper[24928]: I1205 10:54:34.799364 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Dec 05 10:54:34.799926 master-0 kubenswrapper[24928]: I1205 10:54:34.799895 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Dec 05 10:54:34.800692 master-0 kubenswrapper[24928]: I1205 10:54:34.800660 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-6g11pfb8cu15s" Dec 05 10:54:34.800768 master-0 kubenswrapper[24928]: I1205 10:54:34.800715 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Dec 05 10:54:34.800876 master-0 kubenswrapper[24928]: I1205 10:54:34.800847 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Dec 05 10:54:34.801061 master-0 kubenswrapper[24928]: I1205 10:54:34.801030 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 10:54:34.804834 master-0 kubenswrapper[24928]: I1205 10:54:34.804800 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Dec 05 10:54:34.805002 master-0 kubenswrapper[24928]: I1205 10:54:34.804822 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Dec 05 10:54:34.805303 master-0 kubenswrapper[24928]: I1205 10:54:34.805234 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 10:54:34.806816 master-0 kubenswrapper[24928]: I1205 10:54:34.805929 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Dec 05 10:54:34.806816 master-0 kubenswrapper[24928]: I1205 10:54:34.805996 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Dec 05 10:54:34.806816 master-0 kubenswrapper[24928]: I1205 10:54:34.806369 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Dec 05 10:54:34.809718 master-0 kubenswrapper[24928]: I1205 10:54:34.807373 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Dec 05 10:54:34.809718 master-0 kubenswrapper[24928]: I1205 10:54:34.807725 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Dec 05 10:54:34.809718 master-0 kubenswrapper[24928]: I1205 10:54:34.807907 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-71gj50g3moc9k" Dec 05 10:54:34.811195 master-0 kubenswrapper[24928]: I1205 10:54:34.811153 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 10:54:34.812791 master-0 kubenswrapper[24928]: I1205 10:54:34.812733 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Dec 05 10:54:34.842755 master-0 kubenswrapper[24928]: I1205 10:54:34.842707 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 10:54:34.848942 master-0 kubenswrapper[24928]: I1205 10:54:34.848898 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.848942 master-0 kubenswrapper[24928]: I1205 10:54:34.848940 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.849148 master-0 kubenswrapper[24928]: I1205 10:54:34.848959 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-trusted-ca-bundle\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.849148 master-0 kubenswrapper[24928]: I1205 10:54:34.848975 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-audit-log\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.849148 master-0 kubenswrapper[24928]: I1205 10:54:34.849011 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-service-ca\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.849148 master-0 kubenswrapper[24928]: I1205 10:54:34.849027 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.849148 master-0 kubenswrapper[24928]: I1205 10:54:34.849047 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-secret-metrics-server-tls\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.849307 master-0 kubenswrapper[24928]: I1205 10:54:34.849161 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d54hx\" (UniqueName: \"kubernetes.io/projected/25d480c2-7108-450d-adc7-c56e0b5be0c8-kube-api-access-d54hx\") pod \"community-operators-mcjzc\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:34.849345 master-0 kubenswrapper[24928]: I1205 10:54:34.849295 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.849392 master-0 kubenswrapper[24928]: I1205 10:54:34.849374 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-grpc-tls\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.849485 master-0 kubenswrapper[24928]: I1205 10:54:34.849454 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.849538 master-0 kubenswrapper[24928]: I1205 10:54:34.849502 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9dd433a3-fd53-4517-bb7b-84b1457ceb10-serving-cert\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.849627 master-0 kubenswrapper[24928]: I1205 10:54:34.849598 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9dd433a3-fd53-4517-bb7b-84b1457ceb10-client-ca\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.849671 master-0 kubenswrapper[24928]: I1205 10:54:34.849650 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-catalog-content\") pod \"redhat-operators-8pb58\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:34.849725 master-0 kubenswrapper[24928]: I1205 10:54:34.849693 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.849778 master-0 kubenswrapper[24928]: I1205 10:54:34.849743 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-console-config\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.849778 master-0 kubenswrapper[24928]: I1205 10:54:34.849764 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-tls\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.849872 master-0 kubenswrapper[24928]: I1205 10:54:34.849822 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-config\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.849872 master-0 kubenswrapper[24928]: I1205 10:54:34.849870 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.849965 master-0 kubenswrapper[24928]: I1205 10:54:34.849892 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rtlf\" (UniqueName: \"kubernetes.io/projected/97818d19-9a0b-40a3-a075-a3ed60eebe62-kube-api-access-7rtlf\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.849965 master-0 kubenswrapper[24928]: I1205 10:54:34.849913 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/400f4f65-216c-496e-b083-ecc5c11f3875-client-ca\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.849965 master-0 kubenswrapper[24928]: I1205 10:54:34.849934 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.849965 master-0 kubenswrapper[24928]: I1205 10:54:34.849960 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/880060c7-79ed-428b-be59-ff663617615b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.850085 master-0 kubenswrapper[24928]: I1205 10:54:34.849989 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.850085 master-0 kubenswrapper[24928]: I1205 10:54:34.850019 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.850085 master-0 kubenswrapper[24928]: I1205 10:54:34.850043 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-web-config\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.850176 master-0 kubenswrapper[24928]: I1205 10:54:34.850088 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.850242 master-0 kubenswrapper[24928]: I1205 10:54:34.850178 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-serving-cert\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.850347 master-0 kubenswrapper[24928]: I1205 10:54:34.850299 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2ttf\" (UniqueName: \"kubernetes.io/projected/e61eff80-d16c-4636-837d-09237981ca7f-kube-api-access-d2ttf\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.850347 master-0 kubenswrapper[24928]: I1205 10:54:34.850334 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-secret-metrics-client-certs\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.850437 master-0 kubenswrapper[24928]: I1205 10:54:34.850367 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.850482 master-0 kubenswrapper[24928]: I1205 10:54:34.850415 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-config-volume\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.850482 master-0 kubenswrapper[24928]: I1205 10:54:34.850466 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-oauth-serving-cert\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.850539 master-0 kubenswrapper[24928]: I1205 10:54:34.850488 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-utilities\") pod \"redhat-operators-8pb58\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:34.850539 master-0 kubenswrapper[24928]: I1205 10:54:34.850509 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.850539 master-0 kubenswrapper[24928]: I1205 10:54:34.850528 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e61eff80-d16c-4636-837d-09237981ca7f-audit-dir\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.850631 master-0 kubenswrapper[24928]: I1205 10:54:34.850551 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9dd433a3-fd53-4517-bb7b-84b1457ceb10-proxy-ca-bundles\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.850631 master-0 kubenswrapper[24928]: I1205 10:54:34.850571 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/400f4f65-216c-496e-b083-ecc5c11f3875-serving-cert\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.850631 master-0 kubenswrapper[24928]: I1205 10:54:34.850589 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-web-config\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.850631 master-0 kubenswrapper[24928]: I1205 10:54:34.850605 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-utilities\") pod \"certified-operators-52wjg\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:34.850631 master-0 kubenswrapper[24928]: I1205 10:54:34.850621 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-metrics-server-audit-profiles\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.850781 master-0 kubenswrapper[24928]: I1205 10:54:34.850639 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8v4c\" (UniqueName: \"kubernetes.io/projected/400f4f65-216c-496e-b083-ecc5c11f3875-kube-api-access-s8v4c\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.850781 master-0 kubenswrapper[24928]: I1205 10:54:34.850696 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/400f4f65-216c-496e-b083-ecc5c11f3875-config\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.850781 master-0 kubenswrapper[24928]: I1205 10:54:34.850733 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.851206 master-0 kubenswrapper[24928]: I1205 10:54:34.850820 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.851206 master-0 kubenswrapper[24928]: I1205 10:54:34.850861 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brpgb\" (UniqueName: \"kubernetes.io/projected/880060c7-79ed-428b-be59-ff663617615b-kube-api-access-brpgb\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.851206 master-0 kubenswrapper[24928]: I1205 10:54:34.850889 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/97818d19-9a0b-40a3-a075-a3ed60eebe62-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.851206 master-0 kubenswrapper[24928]: I1205 10:54:34.850912 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twrt5\" (UniqueName: \"kubernetes.io/projected/9dd433a3-fd53-4517-bb7b-84b1457ceb10-kube-api-access-twrt5\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.851206 master-0 kubenswrapper[24928]: I1205 10:54:34.850962 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr7nj\" (UniqueName: \"kubernetes.io/projected/2934463a-d494-4735-a5d8-ad2c16aea065-kube-api-access-xr7nj\") pod \"redhat-marketplace-l4grl\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:34.851206 master-0 kubenswrapper[24928]: I1205 10:54:34.851041 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.851206 master-0 kubenswrapper[24928]: I1205 10:54:34.851131 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-client-ca-bundle\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.851206 master-0 kubenswrapper[24928]: I1205 10:54:34.851178 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.851449 master-0 kubenswrapper[24928]: I1205 10:54:34.851224 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-user-template-login\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.851449 master-0 kubenswrapper[24928]: I1205 10:54:34.851291 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.851449 master-0 kubenswrapper[24928]: I1205 10:54:34.851365 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srz7b\" (UniqueName: \"kubernetes.io/projected/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-kube-api-access-srz7b\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.851449 master-0 kubenswrapper[24928]: I1205 10:54:34.851408 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-user-template-error\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.851607 master-0 kubenswrapper[24928]: I1205 10:54:34.851468 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/97818d19-9a0b-40a3-a075-a3ed60eebe62-config-out\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.851607 master-0 kubenswrapper[24928]: I1205 10:54:34.851518 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-metrics-client-ca\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.851670 master-0 kubenswrapper[24928]: I1205 10:54:34.851613 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-catalog-content\") pod \"certified-operators-52wjg\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:34.851670 master-0 kubenswrapper[24928]: I1205 10:54:34.851651 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-oauth-config\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.851726 master-0 kubenswrapper[24928]: I1205 10:54:34.851704 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/880060c7-79ed-428b-be59-ff663617615b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.851793 master-0 kubenswrapper[24928]: I1205 10:54:34.851758 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-audit-policies\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.851831 master-0 kubenswrapper[24928]: I1205 10:54:34.851806 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6474b\" (UniqueName: \"kubernetes.io/projected/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-kube-api-access-6474b\") pod \"certified-operators-52wjg\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:34.851873 master-0 kubenswrapper[24928]: I1205 10:54:34.851850 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c2j5\" (UniqueName: \"kubernetes.io/projected/8a54075d-4cba-439e-96c0-db16e991d658-kube-api-access-5c2j5\") pod \"redhat-operators-8pb58\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:34.851906 master-0 kubenswrapper[24928]: I1205 10:54:34.851888 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds24b\" (UniqueName: \"kubernetes.io/projected/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-kube-api-access-ds24b\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.851948 master-0 kubenswrapper[24928]: I1205 10:54:34.851926 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-router-certs\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.851983 master-0 kubenswrapper[24928]: I1205 10:54:34.851962 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-utilities\") pod \"community-operators-mcjzc\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:34.852017 master-0 kubenswrapper[24928]: I1205 10:54:34.851999 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw5lx\" (UniqueName: \"kubernetes.io/projected/629d199b-36e5-4c96-b58f-24314c4f2f58-kube-api-access-tw5lx\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.852098 master-0 kubenswrapper[24928]: I1205 10:54:34.852071 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/880060c7-79ed-428b-be59-ff663617615b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.852144 master-0 kubenswrapper[24928]: I1205 10:54:34.852121 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/880060c7-79ed-428b-be59-ff663617615b-config-out\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.852191 master-0 kubenswrapper[24928]: I1205 10:54:34.852149 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-catalog-content\") pod \"community-operators-mcjzc\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:34.852241 master-0 kubenswrapper[24928]: I1205 10:54:34.852189 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/97818d19-9a0b-40a3-a075-a3ed60eebe62-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.852241 master-0 kubenswrapper[24928]: I1205 10:54:34.852235 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.852341 master-0 kubenswrapper[24928]: I1205 10:54:34.852273 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.852341 master-0 kubenswrapper[24928]: I1205 10:54:34.852307 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.852408 master-0 kubenswrapper[24928]: I1205 10:54:34.852349 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.852408 master-0 kubenswrapper[24928]: I1205 10:54:34.852383 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.852518 master-0 kubenswrapper[24928]: I1205 10:54:34.852489 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-catalog-content\") pod \"redhat-marketplace-l4grl\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:34.852572 master-0 kubenswrapper[24928]: I1205 10:54:34.852553 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dd433a3-fd53-4517-bb7b-84b1457ceb10-config\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.852613 master-0 kubenswrapper[24928]: I1205 10:54:34.852594 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-service-ca\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.852645 master-0 kubenswrapper[24928]: I1205 10:54:34.852634 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.852829 master-0 kubenswrapper[24928]: I1205 10:54:34.852792 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.852869 master-0 kubenswrapper[24928]: I1205 10:54:34.852838 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-utilities\") pod \"redhat-marketplace-l4grl\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:34.852902 master-0 kubenswrapper[24928]: I1205 10:54:34.852880 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.853028 master-0 kubenswrapper[24928]: I1205 10:54:34.852991 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/880060c7-79ed-428b-be59-ff663617615b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.853084 master-0 kubenswrapper[24928]: I1205 10:54:34.853061 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-session\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.870949 master-0 kubenswrapper[24928]: I1205 10:54:34.870895 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 10:54:34.875348 master-0 kubenswrapper[24928]: I1205 10:54:34.875305 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 10:54:34.924170 master-0 kubenswrapper[24928]: I1205 10:54:34.924120 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 10:54:34.954473 master-0 kubenswrapper[24928]: I1205 10:54:34.954358 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.954473 master-0 kubenswrapper[24928]: I1205 10:54:34.954466 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srz7b\" (UniqueName: \"kubernetes.io/projected/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-kube-api-access-srz7b\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.955683 master-0 kubenswrapper[24928]: I1205 10:54:34.954525 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-user-template-error\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.955683 master-0 kubenswrapper[24928]: I1205 10:54:34.954709 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/97818d19-9a0b-40a3-a075-a3ed60eebe62-config-out\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.955683 master-0 kubenswrapper[24928]: I1205 10:54:34.954762 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-metrics-client-ca\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.955683 master-0 kubenswrapper[24928]: I1205 10:54:34.954788 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-catalog-content\") pod \"certified-operators-52wjg\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:34.955683 master-0 kubenswrapper[24928]: I1205 10:54:34.954807 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-oauth-config\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.955683 master-0 kubenswrapper[24928]: I1205 10:54:34.954962 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/880060c7-79ed-428b-be59-ff663617615b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.955683 master-0 kubenswrapper[24928]: I1205 10:54:34.955255 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-audit-policies\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.955683 master-0 kubenswrapper[24928]: I1205 10:54:34.955322 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6474b\" (UniqueName: \"kubernetes.io/projected/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-kube-api-access-6474b\") pod \"certified-operators-52wjg\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:34.955683 master-0 kubenswrapper[24928]: I1205 10:54:34.955582 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c2j5\" (UniqueName: \"kubernetes.io/projected/8a54075d-4cba-439e-96c0-db16e991d658-kube-api-access-5c2j5\") pod \"redhat-operators-8pb58\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:34.955683 master-0 kubenswrapper[24928]: I1205 10:54:34.955623 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-catalog-content\") pod \"certified-operators-52wjg\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:34.955683 master-0 kubenswrapper[24928]: I1205 10:54:34.955662 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds24b\" (UniqueName: \"kubernetes.io/projected/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-kube-api-access-ds24b\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.955720 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-router-certs\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.955894 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-utilities\") pod \"community-operators-mcjzc\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.955934 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw5lx\" (UniqueName: \"kubernetes.io/projected/629d199b-36e5-4c96-b58f-24314c4f2f58-kube-api-access-tw5lx\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.955962 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/880060c7-79ed-428b-be59-ff663617615b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956052 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/880060c7-79ed-428b-be59-ff663617615b-config-out\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956079 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-catalog-content\") pod \"community-operators-mcjzc\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956103 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/97818d19-9a0b-40a3-a075-a3ed60eebe62-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956128 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956157 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956184 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956206 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956223 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956253 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-catalog-content\") pod \"redhat-marketplace-l4grl\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956303 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dd433a3-fd53-4517-bb7b-84b1457ceb10-config\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956323 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-service-ca\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956347 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956365 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956389 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-utilities\") pod \"redhat-marketplace-l4grl\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956413 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956487 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/880060c7-79ed-428b-be59-ff663617615b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956514 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-session\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956553 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956582 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956600 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-trusted-ca-bundle\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956617 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-audit-log\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.956644 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-service-ca\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.957024 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/880060c7-79ed-428b-be59-ff663617615b-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.957697 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-audit-policies\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.957744 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/880060c7-79ed-428b-be59-ff663617615b-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.958131 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.958160 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.958582 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-utilities\") pod \"community-operators-mcjzc\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.958624 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-metrics-client-ca\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.958654 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/97818d19-9a0b-40a3-a075-a3ed60eebe62-config-out\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.958836 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9dd433a3-fd53-4517-bb7b-84b1457ceb10-config\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.958901 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.958949 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-secret-metrics-server-tls\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.958955 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-service-ca\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.958950 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-utilities\") pod \"redhat-marketplace-l4grl\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.958977 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d54hx\" (UniqueName: \"kubernetes.io/projected/25d480c2-7108-450d-adc7-c56e0b5be0c8-kube-api-access-d54hx\") pod \"community-operators-mcjzc\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.959029 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.959057 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-grpc-tls\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.959090 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.959092 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.959238 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-service-ca\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.959758 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-catalog-content\") pod \"community-operators-mcjzc\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.959778 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-audit-log\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.959925 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9dd433a3-fd53-4517-bb7b-84b1457ceb10-serving-cert\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.959970 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9dd433a3-fd53-4517-bb7b-84b1457ceb10-client-ca\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.960001 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-catalog-content\") pod \"redhat-operators-8pb58\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.960026 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.960038 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-catalog-content\") pod \"redhat-marketplace-l4grl\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.960055 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-console-config\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.960083 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-tls\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.960111 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-config\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.960495 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/880060c7-79ed-428b-be59-ff663617615b-tls-assets\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.960614 master-0 kubenswrapper[24928]: I1205 10:54:34.960740 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-catalog-content\") pod \"redhat-operators-8pb58\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.960771 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-user-template-error\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.960806 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-trusted-ca-bundle\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.960964 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.962454 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.962484 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-router-certs\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.962535 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-console-config\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.962612 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.963595 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.963960 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/400f4f65-216c-496e-b083-ecc5c11f3875-client-ca\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964011 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rtlf\" (UniqueName: \"kubernetes.io/projected/97818d19-9a0b-40a3-a075-a3ed60eebe62-kube-api-access-7rtlf\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964054 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/880060c7-79ed-428b-be59-ff663617615b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964091 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9dd433a3-fd53-4517-bb7b-84b1457ceb10-serving-cert\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964111 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964154 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964280 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964325 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-oauth-config\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964397 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964447 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964699 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9dd433a3-fd53-4517-bb7b-84b1457ceb10-client-ca\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964743 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-web-config\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964801 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964820 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-serving-cert\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964878 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2ttf\" (UniqueName: \"kubernetes.io/projected/e61eff80-d16c-4636-837d-09237981ca7f-kube-api-access-d2ttf\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964901 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-secret-metrics-client-certs\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964924 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964944 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-config-volume\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964963 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-oauth-serving-cert\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.964983 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-utilities\") pod \"redhat-operators-8pb58\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965002 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965024 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e61eff80-d16c-4636-837d-09237981ca7f-audit-dir\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965045 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9dd433a3-fd53-4517-bb7b-84b1457ceb10-proxy-ca-bundles\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965064 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/400f4f65-216c-496e-b083-ecc5c11f3875-serving-cert\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965084 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-web-config\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965103 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-utilities\") pod \"certified-operators-52wjg\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965120 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-metrics-server-audit-profiles\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965137 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8v4c\" (UniqueName: \"kubernetes.io/projected/400f4f65-216c-496e-b083-ecc5c11f3875-kube-api-access-s8v4c\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965160 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/400f4f65-216c-496e-b083-ecc5c11f3875-config\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965177 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965180 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/880060c7-79ed-428b-be59-ff663617615b-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965198 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twrt5\" (UniqueName: \"kubernetes.io/projected/9dd433a3-fd53-4517-bb7b-84b1457ceb10-kube-api-access-twrt5\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965305 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965349 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/400f4f65-216c-496e-b083-ecc5c11f3875-client-ca\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965364 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brpgb\" (UniqueName: \"kubernetes.io/projected/880060c7-79ed-428b-be59-ff663617615b-kube-api-access-brpgb\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965462 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/97818d19-9a0b-40a3-a075-a3ed60eebe62-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965514 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr7nj\" (UniqueName: \"kubernetes.io/projected/2934463a-d494-4735-a5d8-ad2c16aea065-kube-api-access-xr7nj\") pod \"redhat-marketplace-l4grl\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965555 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965606 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-client-ca-bundle\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965645 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.965693 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-user-template-login\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.966294 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e61eff80-d16c-4636-837d-09237981ca7f-audit-dir\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.966299 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-tls\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.966524 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/97818d19-9a0b-40a3-a075-a3ed60eebe62-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.966777 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.967041 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.967267 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-oauth-serving-cert\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.967339 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/97818d19-9a0b-40a3-a075-a3ed60eebe62-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.967432 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-utilities\") pod \"redhat-operators-8pb58\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:34.967688 master-0 kubenswrapper[24928]: I1205 10:54:34.967462 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-utilities\") pod \"certified-operators-52wjg\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.968170 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-secret-metrics-server-tls\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.968334 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.968589 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.968733 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/400f4f65-216c-496e-b083-ecc5c11f3875-config\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.970249 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-metrics-server-audit-profiles\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.971046 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-web-config\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.971148 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9dd433a3-fd53-4517-bb7b-84b1457ceb10-proxy-ca-bundles\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.971519 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.972173 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/97818d19-9a0b-40a3-a075-a3ed60eebe62-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.972962 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.973443 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.973401 master-0 kubenswrapper[24928]: I1205 10:54:34.973491 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-user-template-login\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.974798 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.974886 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-web-config\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.974907 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/880060c7-79ed-428b-be59-ff663617615b-config-out\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.974909 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/400f4f65-216c-496e-b083-ecc5c11f3875-serving-cert\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.975007 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.975089 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-secret-metrics-client-certs\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.975137 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-grpc-tls\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.975186 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.975392 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.975452 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.975415 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e61eff80-d16c-4636-837d-09237981ca7f-v4-0-config-system-session\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.975803 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.976034 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-config-volume\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.976117 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/880060c7-79ed-428b-be59-ff663617615b-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.976671 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.976876 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.977005 master-0 kubenswrapper[24928]: I1205 10:54:34.976955 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/97818d19-9a0b-40a3-a075-a3ed60eebe62-config\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:34.978107 master-0 kubenswrapper[24928]: I1205 10:54:34.977043 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-serving-cert\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:34.978107 master-0 kubenswrapper[24928]: I1205 10:54:34.977224 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.978107 master-0 kubenswrapper[24928]: I1205 10:54:34.977279 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-client-ca-bundle\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:34.978472 master-0 kubenswrapper[24928]: I1205 10:54:34.978434 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:34.997470 master-0 kubenswrapper[24928]: I1205 10:54:34.997184 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-5972r" Dec 05 10:54:35.008013 master-0 kubenswrapper[24928]: I1205 10:54:35.006226 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 10:54:35.063980 master-0 kubenswrapper[24928]: I1205 10:54:35.063893 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 10:54:35.202360 master-0 kubenswrapper[24928]: I1205 10:54:35.153325 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8f9584d48-fblwk"] Dec 05 10:54:35.202360 master-0 kubenswrapper[24928]: I1205 10:54:35.181209 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rtlf\" (UniqueName: \"kubernetes.io/projected/97818d19-9a0b-40a3-a075-a3ed60eebe62-kube-api-access-7rtlf\") pod \"prometheus-k8s-0\" (UID: \"97818d19-9a0b-40a3-a075-a3ed60eebe62\") " pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:35.202360 master-0 kubenswrapper[24928]: I1205 10:54:35.182486 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74977ddd8b-dkrkh"] Dec 05 10:54:35.202360 master-0 kubenswrapper[24928]: I1205 10:54:35.186857 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twrt5\" (UniqueName: \"kubernetes.io/projected/9dd433a3-fd53-4517-bb7b-84b1457ceb10-kube-api-access-twrt5\") pod \"controller-manager-8f9584d48-fblwk\" (UID: \"9dd433a3-fd53-4517-bb7b-84b1457ceb10\") " pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:35.202360 master-0 kubenswrapper[24928]: I1205 10:54:35.187658 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:35.202360 master-0 kubenswrapper[24928]: I1205 10:54:35.190868 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr7nj\" (UniqueName: \"kubernetes.io/projected/2934463a-d494-4735-a5d8-ad2c16aea065-kube-api-access-xr7nj\") pod \"redhat-marketplace-l4grl\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:35.202360 master-0 kubenswrapper[24928]: I1205 10:54:35.190951 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv"] Dec 05 10:54:35.202360 master-0 kubenswrapper[24928]: I1205 10:54:35.197351 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srz7b\" (UniqueName: \"kubernetes.io/projected/1b2240bc-13f2-457d-b9b1-72ca9f8035d1-kube-api-access-srz7b\") pod \"metrics-server-64494f74c5-sqgmf\" (UID: \"1b2240bc-13f2-457d-b9b1-72ca9f8035d1\") " pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:35.202360 master-0 kubenswrapper[24928]: I1205 10:54:35.197677 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mcjzc"] Dec 05 10:54:35.204195 master-0 kubenswrapper[24928]: I1205 10:54:35.204143 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2ttf\" (UniqueName: \"kubernetes.io/projected/e61eff80-d16c-4636-837d-09237981ca7f-kube-api-access-d2ttf\") pod \"oauth-openshift-775fbfd4bb-cxrjv\" (UID: \"e61eff80-d16c-4636-837d-09237981ca7f\") " pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:35.206279 master-0 kubenswrapper[24928]: I1205 10:54:35.205581 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-52wjg"] Dec 05 10:54:35.209174 master-0 kubenswrapper[24928]: I1205 10:54:35.206872 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c2j5\" (UniqueName: \"kubernetes.io/projected/8a54075d-4cba-439e-96c0-db16e991d658-kube-api-access-5c2j5\") pod \"redhat-operators-8pb58\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:35.209174 master-0 kubenswrapper[24928]: I1205 10:54:35.208725 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d54hx\" (UniqueName: \"kubernetes.io/projected/25d480c2-7108-450d-adc7-c56e0b5be0c8-kube-api-access-d54hx\") pod \"community-operators-mcjzc\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:35.209174 master-0 kubenswrapper[24928]: I1205 10:54:35.209062 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-scdp6" Dec 05 10:54:35.209174 master-0 kubenswrapper[24928]: I1205 10:54:35.209057 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8v4c\" (UniqueName: \"kubernetes.io/projected/400f4f65-216c-496e-b083-ecc5c11f3875-kube-api-access-s8v4c\") pod \"route-controller-manager-6c646947f8-brjzq\" (UID: \"400f4f65-216c-496e-b083-ecc5c11f3875\") " pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:35.210165 master-0 kubenswrapper[24928]: I1205 10:54:35.210121 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-credential-operator"/"cloud-credential-operator-dockercfg-ftghr" Dec 05 10:54:35.213535 master-0 kubenswrapper[24928]: I1205 10:54:35.213299 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brpgb\" (UniqueName: \"kubernetes.io/projected/880060c7-79ed-428b-be59-ff663617615b-kube-api-access-brpgb\") pod \"alertmanager-main-0\" (UID: \"880060c7-79ed-428b-be59-ff663617615b\") " pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:35.214243 master-0 kubenswrapper[24928]: I1205 10:54:35.214192 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6474b\" (UniqueName: \"kubernetes.io/projected/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-kube-api-access-6474b\") pod \"certified-operators-52wjg\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:35.214646 master-0 kubenswrapper[24928]: I1205 10:54:35.214539 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds24b\" (UniqueName: \"kubernetes.io/projected/c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7-kube-api-access-ds24b\") pod \"thanos-querier-598896584f-9pd95\" (UID: \"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7\") " pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:35.214646 master-0 kubenswrapper[24928]: I1205 10:54:35.214596 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq"] Dec 05 10:54:35.217332 master-0 kubenswrapper[24928]: I1205 10:54:35.217083 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 10:54:35.219303 master-0 kubenswrapper[24928]: I1205 10:54:35.219268 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:35.222641 master-0 kubenswrapper[24928]: I1205 10:54:35.221883 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw5lx\" (UniqueName: \"kubernetes.io/projected/629d199b-36e5-4c96-b58f-24314c4f2f58-kube-api-access-tw5lx\") pod \"console-74977ddd8b-dkrkh\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:35.222641 master-0 kubenswrapper[24928]: I1205 10:54:35.221971 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8pb58"] Dec 05 10:54:35.230458 master-0 kubenswrapper[24928]: I1205 10:54:35.229810 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-64494f74c5-sqgmf"] Dec 05 10:54:35.234753 master-0 kubenswrapper[24928]: I1205 10:54:35.234680 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-598896584f-9pd95"] Dec 05 10:54:35.240312 master-0 kubenswrapper[24928]: I1205 10:54:35.239561 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:35.246798 master-0 kubenswrapper[24928]: I1205 10:54:35.246111 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 05 10:54:35.260682 master-0 kubenswrapper[24928]: I1205 10:54:35.251186 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-86b5fdbff8-6l4nn"] Dec 05 10:54:35.260682 master-0 kubenswrapper[24928]: I1205 10:54:35.252015 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:35.260682 master-0 kubenswrapper[24928]: I1205 10:54:35.252085 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.260682 master-0 kubenswrapper[24928]: I1205 10:54:35.257696 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-m2gng" Dec 05 10:54:35.277355 master-0 kubenswrapper[24928]: I1205 10:54:35.277287 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-console-config\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.277355 master-0 kubenswrapper[24928]: I1205 10:54:35.277347 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-service-ca\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.277676 master-0 kubenswrapper[24928]: I1205 10:54:35.277380 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-oauth-config\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.277676 master-0 kubenswrapper[24928]: I1205 10:54:35.277464 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhptk\" (UniqueName: \"kubernetes.io/projected/be90caa2-11f2-41f3-af4b-2df058459e12-kube-api-access-zhptk\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.277676 master-0 kubenswrapper[24928]: I1205 10:54:35.277503 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-oauth-serving-cert\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.277676 master-0 kubenswrapper[24928]: I1205 10:54:35.277537 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-serving-cert\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.277676 master-0 kubenswrapper[24928]: I1205 10:54:35.277557 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-trusted-ca-bundle\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.282322 master-0 kubenswrapper[24928]: I1205 10:54:35.282252 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86b5fdbff8-6l4nn"] Dec 05 10:54:35.285518 master-0 kubenswrapper[24928]: I1205 10:54:35.284466 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-master-0" podStartSLOduration=22.284441228 podStartE2EDuration="22.284441228s" podCreationTimestamp="2025-12-05 10:54:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:54:35.281343307 +0000 UTC m=+435.284537168" watchObservedRunningTime="2025-12-05 10:54:35.284441228 +0000 UTC m=+435.287635089" Dec 05 10:54:35.288109 master-0 kubenswrapper[24928]: I1205 10:54:35.288010 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 05 10:54:35.294775 master-0 kubenswrapper[24928]: I1205 10:54:35.293574 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4grl"] Dec 05 10:54:35.320709 master-0 kubenswrapper[24928]: I1205 10:54:35.320615 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 10:54:35.332217 master-0 kubenswrapper[24928]: I1205 10:54:35.332048 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podStartSLOduration=21.33202941 podStartE2EDuration="21.33202941s" podCreationTimestamp="2025-12-05 10:54:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:54:35.328253572 +0000 UTC m=+435.331447443" watchObservedRunningTime="2025-12-05 10:54:35.33202941 +0000 UTC m=+435.335223261" Dec 05 10:54:35.349703 master-0 kubenswrapper[24928]: I1205 10:54:35.348980 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:35.367052 master-0 kubenswrapper[24928]: I1205 10:54:35.366631 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Dec 05 10:54:35.373012 master-0 kubenswrapper[24928]: I1205 10:54:35.372956 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Dec 05 10:54:35.389671 master-0 kubenswrapper[24928]: I1205 10:54:35.389627 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:35.390466 master-0 kubenswrapper[24928]: I1205 10:54:35.389981 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:35.391152 master-0 kubenswrapper[24928]: I1205 10:54:35.391115 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-console-config\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.391152 master-0 kubenswrapper[24928]: I1205 10:54:35.391145 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-service-ca\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.391226 master-0 kubenswrapper[24928]: I1205 10:54:35.391178 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-oauth-config\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.391258 master-0 kubenswrapper[24928]: I1205 10:54:35.391238 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhptk\" (UniqueName: \"kubernetes.io/projected/be90caa2-11f2-41f3-af4b-2df058459e12-kube-api-access-zhptk\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.391298 master-0 kubenswrapper[24928]: I1205 10:54:35.391274 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-oauth-serving-cert\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.391342 master-0 kubenswrapper[24928]: I1205 10:54:35.391307 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-serving-cert\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.391384 master-0 kubenswrapper[24928]: I1205 10:54:35.391338 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-trusted-ca-bundle\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.398452 master-0 kubenswrapper[24928]: I1205 10:54:35.392470 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-trusted-ca-bundle\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.398452 master-0 kubenswrapper[24928]: I1205 10:54:35.393106 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-console-config\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.398452 master-0 kubenswrapper[24928]: I1205 10:54:35.393793 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-service-ca\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.398452 master-0 kubenswrapper[24928]: I1205 10:54:35.398361 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-oauth-serving-cert\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.418816 master-0 kubenswrapper[24928]: I1205 10:54:35.416399 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:35.418816 master-0 kubenswrapper[24928]: I1205 10:54:35.417761 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-serving-cert\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.422270 master-0 kubenswrapper[24928]: I1205 10:54:35.422223 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-oauth-config\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.432628 master-0 kubenswrapper[24928]: I1205 10:54:35.431892 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-node-tuning-operator"/"performance-addon-operator-webhook-cert" Dec 05 10:54:35.432628 master-0 kubenswrapper[24928]: I1205 10:54:35.432245 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:35.436686 master-0 kubenswrapper[24928]: I1205 10:54:35.435631 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhptk\" (UniqueName: \"kubernetes.io/projected/be90caa2-11f2-41f3-af4b-2df058459e12-kube-api-access-zhptk\") pod \"console-86b5fdbff8-6l4nn\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.448631 master-0 kubenswrapper[24928]: I1205 10:54:35.447862 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 10:54:35.459730 master-0 kubenswrapper[24928]: I1205 10:54:35.456952 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:35.468783 master-0 kubenswrapper[24928]: I1205 10:54:35.468738 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-qq76p" Dec 05 10:54:35.475336 master-0 kubenswrapper[24928]: I1205 10:54:35.475277 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 10:54:35.496648 master-0 kubenswrapper[24928]: I1205 10:54:35.496606 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:35.598928 master-0 kubenswrapper[24928]: I1205 10:54:35.598726 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:35.616120 master-0 kubenswrapper[24928]: I1205 10:54:35.616087 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 10:54:35.644768 master-0 kubenswrapper[24928]: I1205 10:54:35.644677 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 10:54:35.680316 master-0 kubenswrapper[24928]: I1205 10:54:35.679943 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-xntfn" Dec 05 10:54:35.738489 master-0 kubenswrapper[24928]: I1205 10:54:35.738294 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Dec 05 10:54:35.808550 master-0 kubenswrapper[24928]: I1205 10:54:35.808166 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 10:54:35.831110 master-0 kubenswrapper[24928]: I1205 10:54:35.830960 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 10:54:35.852970 master-0 kubenswrapper[24928]: I1205 10:54:35.852914 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv"] Dec 05 10:54:35.871772 master-0 kubenswrapper[24928]: W1205 10:54:35.871722 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode61eff80_d16c_4636_837d_09237981ca7f.slice/crio-3697487ff928a3243bd996837a5f75713a204ca484f9434b897ef6244b8d982c WatchSource:0}: Error finding container 3697487ff928a3243bd996837a5f75713a204ca484f9434b897ef6244b8d982c: Status 404 returned error can't find the container with id 3697487ff928a3243bd996837a5f75713a204ca484f9434b897ef6244b8d982c Dec 05 10:54:35.882303 master-0 kubenswrapper[24928]: I1205 10:54:35.882257 24928 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0"] Dec 05 10:54:35.882663 master-0 kubenswrapper[24928]: I1205 10:54:35.882538 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" podUID="b83ccd6fa217a93a2c607d0109896ef8" containerName="startup-monitor" containerID="cri-o://b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38" gracePeriod=5 Dec 05 10:54:35.884223 master-0 kubenswrapper[24928]: I1205 10:54:35.884196 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 10:54:35.899626 master-0 kubenswrapper[24928]: I1205 10:54:35.898658 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 10:54:35.925715 master-0 kubenswrapper[24928]: I1205 10:54:35.924993 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8f9584d48-fblwk"] Dec 05 10:54:35.931208 master-0 kubenswrapper[24928]: W1205 10:54:35.931169 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dd433a3_fd53_4517_bb7b_84b1457ceb10.slice/crio-ed3a6edd6c6f72aa8a5ae9baad7aeb8e46b37a8014c36b70f9542594320d2b8c WatchSource:0}: Error finding container ed3a6edd6c6f72aa8a5ae9baad7aeb8e46b37a8014c36b70f9542594320d2b8c: Status 404 returned error can't find the container with id ed3a6edd6c6f72aa8a5ae9baad7aeb8e46b37a8014c36b70f9542594320d2b8c Dec 05 10:54:36.003587 master-0 kubenswrapper[24928]: I1205 10:54:36.003480 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 10:54:36.028763 master-0 kubenswrapper[24928]: I1205 10:54:36.028720 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-controller-manager-operator"/"cloud-controller-manager-operator-tls" Dec 05 10:54:36.061417 master-0 kubenswrapper[24928]: I1205 10:54:36.061366 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 10:54:36.069240 master-0 kubenswrapper[24928]: I1205 10:54:36.069196 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 10:54:36.085289 master-0 kubenswrapper[24928]: I1205 10:54:36.085230 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 10:54:36.149887 master-0 kubenswrapper[24928]: I1205 10:54:36.149225 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Dec 05 10:54:36.215663 master-0 kubenswrapper[24928]: I1205 10:54:36.215600 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527" path="/var/lib/kubelet/pods/3cf0f61e-27cf-4f7d-b8da-c2f8f88f6527/volumes" Dec 05 10:54:36.296001 master-0 kubenswrapper[24928]: I1205 10:54:36.295926 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 10:54:36.336174 master-0 kubenswrapper[24928]: I1205 10:54:36.335990 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 10:54:36.352471 master-0 kubenswrapper[24928]: I1205 10:54:36.352370 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 10:54:36.352655 master-0 kubenswrapper[24928]: I1205 10:54:36.352645 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 10:54:36.378399 master-0 kubenswrapper[24928]: I1205 10:54:36.378292 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-74977ddd8b-dkrkh"] Dec 05 10:54:36.383064 master-0 kubenswrapper[24928]: W1205 10:54:36.382565 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod629d199b_36e5_4c96_b58f_24314c4f2f58.slice/crio-ba2dee4c8f6ea28a581d1b8366715076c17c0d605303d3209026cf7512384aca WatchSource:0}: Error finding container ba2dee4c8f6ea28a581d1b8366715076c17c0d605303d3209026cf7512384aca: Status 404 returned error can't find the container with id ba2dee4c8f6ea28a581d1b8366715076c17c0d605303d3209026cf7512384aca Dec 05 10:54:36.384805 master-0 kubenswrapper[24928]: W1205 10:54:36.384752 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2934463a_d494_4735_a5d8_ad2c16aea065.slice/crio-1f39e81a5d922f5c8f7636d4fc0f9b6b55b79a2ac69769e2833a8169baa5278c WatchSource:0}: Error finding container 1f39e81a5d922f5c8f7636d4fc0f9b6b55b79a2ac69769e2833a8169baa5278c: Status 404 returned error can't find the container with id 1f39e81a5d922f5c8f7636d4fc0f9b6b55b79a2ac69769e2833a8169baa5278c Dec 05 10:54:36.385945 master-0 kubenswrapper[24928]: I1205 10:54:36.385905 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq"] Dec 05 10:54:36.394796 master-0 kubenswrapper[24928]: I1205 10:54:36.392905 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4grl"] Dec 05 10:54:36.407654 master-0 kubenswrapper[24928]: I1205 10:54:36.407586 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-598896584f-9pd95"] Dec 05 10:54:36.428633 master-0 kubenswrapper[24928]: W1205 10:54:36.428594 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc4b24b10_85c9_4ef1_8cd3_3afc0cc65af7.slice/crio-faa7fcc6ee218a6d9473827e04893ab9de35f408af27a528228d9544f4560a10 WatchSource:0}: Error finding container faa7fcc6ee218a6d9473827e04893ab9de35f408af27a528228d9544f4560a10: Status 404 returned error can't find the container with id faa7fcc6ee218a6d9473827e04893ab9de35f408af27a528228d9544f4560a10 Dec 05 10:54:36.430126 master-0 kubenswrapper[24928]: I1205 10:54:36.430059 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Dec 05 10:54:36.431582 master-0 kubenswrapper[24928]: I1205 10:54:36.431358 24928 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:54:36.482336 master-0 kubenswrapper[24928]: I1205 10:54:36.482298 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 10:54:36.496859 master-0 kubenswrapper[24928]: I1205 10:54:36.496652 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Dec 05 10:54:36.500913 master-0 kubenswrapper[24928]: I1205 10:54:36.499947 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"880060c7-79ed-428b-be59-ff663617615b","Type":"ContainerStarted","Data":"577a5e27486db1f90303a668ac866f03a91aada70775d8a5bf07c295b55d755e"} Dec 05 10:54:36.500913 master-0 kubenswrapper[24928]: I1205 10:54:36.500892 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4grl" event={"ID":"2934463a-d494-4735-a5d8-ad2c16aea065","Type":"ContainerStarted","Data":"1f39e81a5d922f5c8f7636d4fc0f9b6b55b79a2ac69769e2833a8169baa5278c"} Dec 05 10:54:36.506008 master-0 kubenswrapper[24928]: I1205 10:54:36.505905 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-olm-operator"/"openshift-service-ca.crt" Dec 05 10:54:36.509414 master-0 kubenswrapper[24928]: I1205 10:54:36.509351 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" event={"ID":"400f4f65-216c-496e-b083-ecc5c11f3875","Type":"ContainerStarted","Data":"f3884b53f283ec1058cc88d27c6c76baed431cf203d72387025a76c29948177c"} Dec 05 10:54:36.518168 master-0 kubenswrapper[24928]: I1205 10:54:36.518002 24928 generic.go:334] "Generic (PLEG): container finished" podID="97818d19-9a0b-40a3-a075-a3ed60eebe62" containerID="b283dc3403c896996efcd24d2b435929e44cfd40f4f6a4e091a1839fd63f20d9" exitCode=0 Dec 05 10:54:36.518168 master-0 kubenswrapper[24928]: I1205 10:54:36.518087 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"97818d19-9a0b-40a3-a075-a3ed60eebe62","Type":"ContainerDied","Data":"b283dc3403c896996efcd24d2b435929e44cfd40f4f6a4e091a1839fd63f20d9"} Dec 05 10:54:36.518168 master-0 kubenswrapper[24928]: I1205 10:54:36.518118 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"97818d19-9a0b-40a3-a075-a3ed60eebe62","Type":"ContainerStarted","Data":"e422a43c0ca5f7742bd2b531617d464fb02a5d60589407579d742d1a80e156d9"} Dec 05 10:54:36.521345 master-0 kubenswrapper[24928]: I1205 10:54:36.520371 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 10:54:36.521345 master-0 kubenswrapper[24928]: I1205 10:54:36.521022 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74977ddd8b-dkrkh" event={"ID":"629d199b-36e5-4c96-b58f-24314c4f2f58","Type":"ContainerStarted","Data":"ba2dee4c8f6ea28a581d1b8366715076c17c0d605303d3209026cf7512384aca"} Dec 05 10:54:36.526124 master-0 kubenswrapper[24928]: I1205 10:54:36.525433 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" event={"ID":"e61eff80-d16c-4636-837d-09237981ca7f","Type":"ContainerStarted","Data":"b351a9e66532abf32a67be0f5a00d3501a653850d257b7b1d3a273f0dd50b411"} Dec 05 10:54:36.526124 master-0 kubenswrapper[24928]: I1205 10:54:36.525491 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" event={"ID":"e61eff80-d16c-4636-837d-09237981ca7f","Type":"ContainerStarted","Data":"3697487ff928a3243bd996837a5f75713a204ca484f9434b897ef6244b8d982c"} Dec 05 10:54:36.526124 master-0 kubenswrapper[24928]: I1205 10:54:36.525871 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:36.528406 master-0 kubenswrapper[24928]: I1205 10:54:36.528325 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" event={"ID":"9dd433a3-fd53-4517-bb7b-84b1457ceb10","Type":"ContainerStarted","Data":"b61022c6134a985fdca43e83a12a938f9750fe162b8a3b0b1692324473749468"} Dec 05 10:54:36.528743 master-0 kubenswrapper[24928]: I1205 10:54:36.528656 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:36.528743 master-0 kubenswrapper[24928]: I1205 10:54:36.528689 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" event={"ID":"9dd433a3-fd53-4517-bb7b-84b1457ceb10","Type":"ContainerStarted","Data":"ed3a6edd6c6f72aa8a5ae9baad7aeb8e46b37a8014c36b70f9542594320d2b8c"} Dec 05 10:54:36.530248 master-0 kubenswrapper[24928]: I1205 10:54:36.530038 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" event={"ID":"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7","Type":"ContainerStarted","Data":"faa7fcc6ee218a6d9473827e04893ab9de35f408af27a528228d9544f4560a10"} Dec 05 10:54:36.535297 master-0 kubenswrapper[24928]: I1205 10:54:36.533148 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" Dec 05 10:54:36.538025 master-0 kubenswrapper[24928]: I1205 10:54:36.537108 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 10:54:36.560718 master-0 kubenswrapper[24928]: I1205 10:54:36.560330 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" Dec 05 10:54:36.562905 master-0 kubenswrapper[24928]: I1205 10:54:36.562864 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Dec 05 10:54:36.588625 master-0 kubenswrapper[24928]: I1205 10:54:36.588544 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-775fbfd4bb-cxrjv" podStartSLOduration=62.588523945 podStartE2EDuration="1m2.588523945s" podCreationTimestamp="2025-12-05 10:53:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:54:36.578326089 +0000 UTC m=+436.581519950" watchObservedRunningTime="2025-12-05 10:54:36.588523945 +0000 UTC m=+436.591717816" Dec 05 10:54:36.612457 master-0 kubenswrapper[24928]: I1205 10:54:36.610997 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-8f9584d48-fblwk" podStartSLOduration=62.610979525 podStartE2EDuration="1m2.610979525s" podCreationTimestamp="2025-12-05 10:53:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:54:36.608248432 +0000 UTC m=+436.611442303" watchObservedRunningTime="2025-12-05 10:54:36.610979525 +0000 UTC m=+436.614173376" Dec 05 10:54:36.633163 master-0 kubenswrapper[24928]: I1205 10:54:36.633008 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 10:54:36.680480 master-0 kubenswrapper[24928]: I1205 10:54:36.680329 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 10:54:36.683707 master-0 kubenswrapper[24928]: I1205 10:54:36.683585 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 10:54:36.686753 master-0 kubenswrapper[24928]: I1205 10:54:36.686700 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8pb58"] Dec 05 10:54:36.697687 master-0 kubenswrapper[24928]: I1205 10:54:36.697649 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-86b5fdbff8-6l4nn"] Dec 05 10:54:36.727072 master-0 kubenswrapper[24928]: I1205 10:54:36.726756 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mcjzc"] Dec 05 10:54:36.734069 master-0 kubenswrapper[24928]: I1205 10:54:36.734007 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-52wjg"] Dec 05 10:54:36.749582 master-0 kubenswrapper[24928]: I1205 10:54:36.749514 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-64494f74c5-sqgmf"] Dec 05 10:54:36.781366 master-0 kubenswrapper[24928]: I1205 10:54:36.781321 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 10:54:36.864894 master-0 kubenswrapper[24928]: I1205 10:54:36.864849 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 10:54:36.883745 master-0 kubenswrapper[24928]: I1205 10:54:36.882841 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 10:54:36.889904 master-0 kubenswrapper[24928]: I1205 10:54:36.889847 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"trusted-ca" Dec 05 10:54:36.950041 master-0 kubenswrapper[24928]: I1205 10:54:36.948883 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 10:54:36.985541 master-0 kubenswrapper[24928]: I1205 10:54:36.985341 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 10:54:37.009784 master-0 kubenswrapper[24928]: I1205 10:54:37.008048 24928 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 10:54:37.010075 master-0 kubenswrapper[24928]: I1205 10:54:37.010028 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 10:54:37.039437 master-0 kubenswrapper[24928]: I1205 10:54:37.039249 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 10:54:37.116070 master-0 kubenswrapper[24928]: I1205 10:54:37.115904 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"cluster-baremetal-operator-tls" Dec 05 10:54:37.161719 master-0 kubenswrapper[24928]: I1205 10:54:37.160554 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Dec 05 10:54:37.264653 master-0 kubenswrapper[24928]: I1205 10:54:37.264404 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 10:54:37.295516 master-0 kubenswrapper[24928]: I1205 10:54:37.295476 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 10:54:37.313031 master-0 kubenswrapper[24928]: I1205 10:54:37.312937 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 10:54:37.360858 master-0 kubenswrapper[24928]: I1205 10:54:37.360799 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 10:54:37.366292 master-0 kubenswrapper[24928]: I1205 10:54:37.366247 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"service-ca-bundle" Dec 05 10:54:37.377594 master-0 kubenswrapper[24928]: I1205 10:54:37.377499 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Dec 05 10:54:37.404713 master-0 kubenswrapper[24928]: I1205 10:54:37.404322 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 10:54:37.491916 master-0 kubenswrapper[24928]: I1205 10:54:37.491456 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 10:54:37.507462 master-0 kubenswrapper[24928]: I1205 10:54:37.507323 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Dec 05 10:54:37.542010 master-0 kubenswrapper[24928]: I1205 10:54:37.541972 24928 generic.go:334] "Generic (PLEG): container finished" podID="2934463a-d494-4735-a5d8-ad2c16aea065" containerID="b4840415e898c8b5a1f111906d738ae0935737c840e6761d37b4ebefe81f411a" exitCode=0 Dec 05 10:54:37.542216 master-0 kubenswrapper[24928]: I1205 10:54:37.542125 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4grl" event={"ID":"2934463a-d494-4735-a5d8-ad2c16aea065","Type":"ContainerDied","Data":"b4840415e898c8b5a1f111906d738ae0935737c840e6761d37b4ebefe81f411a"} Dec 05 10:54:37.544306 master-0 kubenswrapper[24928]: I1205 10:54:37.544180 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86b5fdbff8-6l4nn" event={"ID":"be90caa2-11f2-41f3-af4b-2df058459e12","Type":"ContainerStarted","Data":"861de06c2bcb98334d269cb4c41d1465a0dcd05c72e52912435a598cd19bf542"} Dec 05 10:54:37.544306 master-0 kubenswrapper[24928]: I1205 10:54:37.544227 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86b5fdbff8-6l4nn" event={"ID":"be90caa2-11f2-41f3-af4b-2df058459e12","Type":"ContainerStarted","Data":"db59570c8a8ddd33f58367b6f7074b7a3ec8bed27ec3277b0d41df81b8b48d2c"} Dec 05 10:54:37.547965 master-0 kubenswrapper[24928]: I1205 10:54:37.547941 24928 generic.go:334] "Generic (PLEG): container finished" podID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" containerID="28abf4df43ff19287f182dc322a79951c1b7d10826c92fa80a92b4b4ba1bff28" exitCode=0 Dec 05 10:54:37.548158 master-0 kubenswrapper[24928]: I1205 10:54:37.548031 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52wjg" event={"ID":"241003b6-4aa6-4ed5-8bdd-46f2d30662fd","Type":"ContainerDied","Data":"28abf4df43ff19287f182dc322a79951c1b7d10826c92fa80a92b4b4ba1bff28"} Dec 05 10:54:37.548158 master-0 kubenswrapper[24928]: I1205 10:54:37.548072 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52wjg" event={"ID":"241003b6-4aa6-4ed5-8bdd-46f2d30662fd","Type":"ContainerStarted","Data":"f5130e9cce93cfe214efea75ca82a1cbc656c3e820221aab863c357c84937203"} Dec 05 10:54:37.550825 master-0 kubenswrapper[24928]: I1205 10:54:37.550792 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" event={"ID":"1b2240bc-13f2-457d-b9b1-72ca9f8035d1","Type":"ContainerStarted","Data":"a494bfab18e00791c3d4df375c40d587cf5cbb485180c806b5566deb6f3b9871"} Dec 05 10:54:37.550825 master-0 kubenswrapper[24928]: I1205 10:54:37.550818 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" event={"ID":"1b2240bc-13f2-457d-b9b1-72ca9f8035d1","Type":"ContainerStarted","Data":"ebe124d308f9b9a9a37be5030ec16b8307b1c2ea79f7f423faa82333cf9b0cad"} Dec 05 10:54:37.555135 master-0 kubenswrapper[24928]: I1205 10:54:37.555085 24928 generic.go:334] "Generic (PLEG): container finished" podID="880060c7-79ed-428b-be59-ff663617615b" containerID="1b66f35d750fc00ded64128c3e1a4cf1dde2aa962c9e8dd15ee564ee0b4a231b" exitCode=0 Dec 05 10:54:37.555196 master-0 kubenswrapper[24928]: I1205 10:54:37.555176 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"880060c7-79ed-428b-be59-ff663617615b","Type":"ContainerDied","Data":"1b66f35d750fc00ded64128c3e1a4cf1dde2aa962c9e8dd15ee564ee0b4a231b"} Dec 05 10:54:37.560679 master-0 kubenswrapper[24928]: I1205 10:54:37.560624 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" event={"ID":"400f4f65-216c-496e-b083-ecc5c11f3875","Type":"ContainerStarted","Data":"e0cf100f8f8fa6c1347e9c99b31aa481ed9d1cdd3fbb91343d4efdf694b6bc5f"} Dec 05 10:54:37.567871 master-0 kubenswrapper[24928]: I1205 10:54:37.563642 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:37.567871 master-0 kubenswrapper[24928]: I1205 10:54:37.566934 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74977ddd8b-dkrkh" event={"ID":"629d199b-36e5-4c96-b58f-24314c4f2f58","Type":"ContainerStarted","Data":"f1a948781cc195931bc78694a8e65f57a216949ec87c08ca993d58ff6b6fbbd1"} Dec 05 10:54:37.573312 master-0 kubenswrapper[24928]: I1205 10:54:37.573272 24928 generic.go:334] "Generic (PLEG): container finished" podID="8a54075d-4cba-439e-96c0-db16e991d658" containerID="0ee7dee1876c3ed1dcc8f819d2ac65a461808f03de47099ca1a137bd407fb7b7" exitCode=0 Dec 05 10:54:37.573393 master-0 kubenswrapper[24928]: I1205 10:54:37.573355 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8pb58" event={"ID":"8a54075d-4cba-439e-96c0-db16e991d658","Type":"ContainerDied","Data":"0ee7dee1876c3ed1dcc8f819d2ac65a461808f03de47099ca1a137bd407fb7b7"} Dec 05 10:54:37.573468 master-0 kubenswrapper[24928]: I1205 10:54:37.573405 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8pb58" event={"ID":"8a54075d-4cba-439e-96c0-db16e991d658","Type":"ContainerStarted","Data":"bace8793d0c45eb8ada78dde17ec3e9d888c36fef2e3c8dc7397b9a47a136c43"} Dec 05 10:54:37.573654 master-0 kubenswrapper[24928]: I1205 10:54:37.573622 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" Dec 05 10:54:37.576741 master-0 kubenswrapper[24928]: I1205 10:54:37.576707 24928 generic.go:334] "Generic (PLEG): container finished" podID="25d480c2-7108-450d-adc7-c56e0b5be0c8" containerID="f42999b81de2c99cb042a2f8cade529d3e0c6dc94e65a8e525d08f68560b9e73" exitCode=0 Dec 05 10:54:37.576981 master-0 kubenswrapper[24928]: I1205 10:54:37.576943 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcjzc" event={"ID":"25d480c2-7108-450d-adc7-c56e0b5be0c8","Type":"ContainerDied","Data":"f42999b81de2c99cb042a2f8cade529d3e0c6dc94e65a8e525d08f68560b9e73"} Dec 05 10:54:37.577050 master-0 kubenswrapper[24928]: I1205 10:54:37.576994 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcjzc" event={"ID":"25d480c2-7108-450d-adc7-c56e0b5be0c8","Type":"ContainerStarted","Data":"f87785d6016322c345dd513ef378b4a11bf62b7475beba1226df7371b9687bcc"} Dec 05 10:54:37.611213 master-0 kubenswrapper[24928]: I1205 10:54:37.611116 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" podStartSLOduration=63.611090274 podStartE2EDuration="1m3.611090274s" podCreationTimestamp="2025-12-05 10:53:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:54:37.603232342 +0000 UTC m=+437.606426193" watchObservedRunningTime="2025-12-05 10:54:37.611090274 +0000 UTC m=+437.614284125" Dec 05 10:54:37.660195 master-0 kubenswrapper[24928]: I1205 10:54:37.659706 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6c646947f8-brjzq" podStartSLOduration=63.659682289 podStartE2EDuration="1m3.659682289s" podCreationTimestamp="2025-12-05 10:53:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:54:37.628938917 +0000 UTC m=+437.632132788" watchObservedRunningTime="2025-12-05 10:54:37.659682289 +0000 UTC m=+437.662876140" Dec 05 10:54:37.700403 master-0 kubenswrapper[24928]: I1205 10:54:37.700295 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-86b5fdbff8-6l4nn" podStartSLOduration=10.700247238 podStartE2EDuration="10.700247238s" podCreationTimestamp="2025-12-05 10:54:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:54:37.688149548 +0000 UTC m=+437.691343399" watchObservedRunningTime="2025-12-05 10:54:37.700247238 +0000 UTC m=+437.703441089" Dec 05 10:54:37.713385 master-0 kubenswrapper[24928]: I1205 10:54:37.713299 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-74977ddd8b-dkrkh" podStartSLOduration=63.71327666 podStartE2EDuration="1m3.71327666s" podCreationTimestamp="2025-12-05 10:53:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:54:37.710993977 +0000 UTC m=+437.714187828" watchObservedRunningTime="2025-12-05 10:54:37.71327666 +0000 UTC m=+437.716470511" Dec 05 10:54:37.715444 master-0 kubenswrapper[24928]: I1205 10:54:37.715382 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 10:54:37.750841 master-0 kubenswrapper[24928]: I1205 10:54:37.750516 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Dec 05 10:54:37.764736 master-0 kubenswrapper[24928]: I1205 10:54:37.764546 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"kube-root-ca.crt" Dec 05 10:54:37.839465 master-0 kubenswrapper[24928]: I1205 10:54:37.839317 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 10:54:38.075639 master-0 kubenswrapper[24928]: I1205 10:54:38.075591 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 10:54:38.140862 master-0 kubenswrapper[24928]: I1205 10:54:38.140797 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemeter-client-serving-certs-ca-bundle" Dec 05 10:54:38.181579 master-0 kubenswrapper[24928]: I1205 10:54:38.181528 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 10:54:38.241449 master-0 kubenswrapper[24928]: I1205 10:54:38.241393 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"openshift-service-ca.crt" Dec 05 10:54:38.310497 master-0 kubenswrapper[24928]: I1205 10:54:38.310297 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 10:54:38.406739 master-0 kubenswrapper[24928]: I1205 10:54:38.406549 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rrnj5" Dec 05 10:54:38.462355 master-0 kubenswrapper[24928]: I1205 10:54:38.462167 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 10:54:38.588009 master-0 kubenswrapper[24928]: I1205 10:54:38.587966 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 10:54:38.593066 master-0 kubenswrapper[24928]: I1205 10:54:38.592879 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 10:54:38.604144 master-0 kubenswrapper[24928]: I1205 10:54:38.603961 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 10:54:39.016030 master-0 kubenswrapper[24928]: I1205 10:54:39.015965 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 10:54:39.093395 master-0 kubenswrapper[24928]: I1205 10:54:39.093323 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 10:54:39.104936 master-0 kubenswrapper[24928]: I1205 10:54:39.104886 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 10:54:39.117119 master-0 kubenswrapper[24928]: I1205 10:54:39.116981 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cloud-controller-manager-operator"/"cluster-cloud-controller-manager-dockercfg-45kcs" Dec 05 10:54:39.200629 master-0 kubenswrapper[24928]: I1205 10:54:39.200556 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 10:54:39.327047 master-0 kubenswrapper[24928]: I1205 10:54:39.326933 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Dec 05 10:54:39.432187 master-0 kubenswrapper[24928]: I1205 10:54:39.432012 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-insights"/"openshift-service-ca.crt" Dec 05 10:54:39.607532 master-0 kubenswrapper[24928]: I1205 10:54:39.606948 24928 generic.go:334] "Generic (PLEG): container finished" podID="25d480c2-7108-450d-adc7-c56e0b5be0c8" containerID="da15035e6482378c271f4a44a0e9cb9b80044a94ce5194227563641492f09aa7" exitCode=0 Dec 05 10:54:39.607532 master-0 kubenswrapper[24928]: I1205 10:54:39.607019 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcjzc" event={"ID":"25d480c2-7108-450d-adc7-c56e0b5be0c8","Type":"ContainerDied","Data":"da15035e6482378c271f4a44a0e9cb9b80044a94ce5194227563641492f09aa7"} Dec 05 10:54:39.611013 master-0 kubenswrapper[24928]: I1205 10:54:39.610973 24928 generic.go:334] "Generic (PLEG): container finished" podID="2934463a-d494-4735-a5d8-ad2c16aea065" containerID="555f9686aee99474ecd7968cbfef0667db2d656f174e224a0a8651e26cd15b97" exitCode=0 Dec 05 10:54:39.611076 master-0 kubenswrapper[24928]: I1205 10:54:39.611031 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4grl" event={"ID":"2934463a-d494-4735-a5d8-ad2c16aea065","Type":"ContainerDied","Data":"555f9686aee99474ecd7968cbfef0667db2d656f174e224a0a8651e26cd15b97"} Dec 05 10:54:39.624140 master-0 kubenswrapper[24928]: I1205 10:54:39.624077 24928 generic.go:334] "Generic (PLEG): container finished" podID="8a54075d-4cba-439e-96c0-db16e991d658" containerID="5328f8133a36f39e3b380fc3497d03e68bbc44cc892da222cfc20c5c6eda1fe7" exitCode=0 Dec 05 10:54:39.625364 master-0 kubenswrapper[24928]: I1205 10:54:39.625236 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8pb58" event={"ID":"8a54075d-4cba-439e-96c0-db16e991d658","Type":"ContainerDied","Data":"5328f8133a36f39e3b380fc3497d03e68bbc44cc892da222cfc20c5c6eda1fe7"} Dec 05 10:54:39.651656 master-0 kubenswrapper[24928]: I1205 10:54:39.651519 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-olm-operator"/"cluster-olm-operator-serving-cert" Dec 05 10:54:39.671204 master-0 kubenswrapper[24928]: I1205 10:54:39.671123 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-catalogd"/"kube-root-ca.crt" Dec 05 10:54:40.242137 master-0 kubenswrapper[24928]: I1205 10:54:40.242080 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 10:54:40.621988 master-0 kubenswrapper[24928]: I1205 10:54:40.621926 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cloud-controller-manager-operator"/"cloud-controller-manager-images" Dec 05 10:54:40.798564 master-0 kubenswrapper[24928]: I1205 10:54:40.798494 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 10:54:40.916963 master-0 kubenswrapper[24928]: I1205 10:54:40.916283 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Dec 05 10:54:41.071508 master-0 kubenswrapper[24928]: I1205 10:54:41.071449 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-node-tuning-operator"/"kube-root-ca.crt" Dec 05 10:54:41.429444 master-0 kubenswrapper[24928]: I1205 10:54:41.429383 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-master-0_b83ccd6fa217a93a2c607d0109896ef8/startup-monitor/0.log" Dec 05 10:54:41.429840 master-0 kubenswrapper[24928]: I1205 10:54:41.429476 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:54:41.565081 master-0 kubenswrapper[24928]: I1205 10:54:41.564365 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-resource-dir\") pod \"b83ccd6fa217a93a2c607d0109896ef8\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " Dec 05 10:54:41.565081 master-0 kubenswrapper[24928]: I1205 10:54:41.564463 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-manifests\") pod \"b83ccd6fa217a93a2c607d0109896ef8\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " Dec 05 10:54:41.565081 master-0 kubenswrapper[24928]: I1205 10:54:41.564510 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-lock\") pod \"b83ccd6fa217a93a2c607d0109896ef8\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " Dec 05 10:54:41.565081 master-0 kubenswrapper[24928]: I1205 10:54:41.564597 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-log\") pod \"b83ccd6fa217a93a2c607d0109896ef8\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " Dec 05 10:54:41.565081 master-0 kubenswrapper[24928]: I1205 10:54:41.564646 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-pod-resource-dir\") pod \"b83ccd6fa217a93a2c607d0109896ef8\" (UID: \"b83ccd6fa217a93a2c607d0109896ef8\") " Dec 05 10:54:41.565081 master-0 kubenswrapper[24928]: I1205 10:54:41.564977 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-manifests" (OuterVolumeSpecName: "manifests") pod "b83ccd6fa217a93a2c607d0109896ef8" (UID: "b83ccd6fa217a93a2c607d0109896ef8"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:54:41.565081 master-0 kubenswrapper[24928]: I1205 10:54:41.565011 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "b83ccd6fa217a93a2c607d0109896ef8" (UID: "b83ccd6fa217a93a2c607d0109896ef8"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:54:41.565081 master-0 kubenswrapper[24928]: I1205 10:54:41.565035 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-lock" (OuterVolumeSpecName: "var-lock") pod "b83ccd6fa217a93a2c607d0109896ef8" (UID: "b83ccd6fa217a93a2c607d0109896ef8"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:54:41.565081 master-0 kubenswrapper[24928]: I1205 10:54:41.565056 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-log" (OuterVolumeSpecName: "var-log") pod "b83ccd6fa217a93a2c607d0109896ef8" (UID: "b83ccd6fa217a93a2c607d0109896ef8"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:54:41.566214 master-0 kubenswrapper[24928]: I1205 10:54:41.566115 24928 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-resource-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:54:41.566214 master-0 kubenswrapper[24928]: I1205 10:54:41.566154 24928 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-manifests\") on node \"master-0\" DevicePath \"\"" Dec 05 10:54:41.566214 master-0 kubenswrapper[24928]: I1205 10:54:41.566165 24928 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-lock\") on node \"master-0\" DevicePath \"\"" Dec 05 10:54:41.566214 master-0 kubenswrapper[24928]: I1205 10:54:41.566175 24928 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-var-log\") on node \"master-0\" DevicePath \"\"" Dec 05 10:54:41.576550 master-0 kubenswrapper[24928]: I1205 10:54:41.576396 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "b83ccd6fa217a93a2c607d0109896ef8" (UID: "b83ccd6fa217a93a2c607d0109896ef8"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:54:41.640231 master-0 kubenswrapper[24928]: I1205 10:54:41.639469 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcjzc" event={"ID":"25d480c2-7108-450d-adc7-c56e0b5be0c8","Type":"ContainerStarted","Data":"1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2"} Dec 05 10:54:41.642805 master-0 kubenswrapper[24928]: I1205 10:54:41.642260 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4grl" event={"ID":"2934463a-d494-4735-a5d8-ad2c16aea065","Type":"ContainerStarted","Data":"2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9"} Dec 05 10:54:41.650446 master-0 kubenswrapper[24928]: I1205 10:54:41.649614 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-master-0_b83ccd6fa217a93a2c607d0109896ef8/startup-monitor/0.log" Dec 05 10:54:41.650446 master-0 kubenswrapper[24928]: I1205 10:54:41.649665 24928 generic.go:334] "Generic (PLEG): container finished" podID="b83ccd6fa217a93a2c607d0109896ef8" containerID="b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38" exitCode=137 Dec 05 10:54:41.650446 master-0 kubenswrapper[24928]: I1205 10:54:41.649724 24928 scope.go:117] "RemoveContainer" containerID="b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38" Dec 05 10:54:41.650446 master-0 kubenswrapper[24928]: I1205 10:54:41.649825 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" Dec 05 10:54:41.659345 master-0 kubenswrapper[24928]: I1205 10:54:41.659283 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" event={"ID":"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7","Type":"ContainerStarted","Data":"3698285f86ec954463c28c1aecde6eae49a109f3b0f4b0ecfc50440d01cb2618"} Dec 05 10:54:41.668265 master-0 kubenswrapper[24928]: I1205 10:54:41.668029 24928 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/b83ccd6fa217a93a2c607d0109896ef8-pod-resource-dir\") on node \"master-0\" DevicePath \"\"" Dec 05 10:54:41.675538 master-0 kubenswrapper[24928]: I1205 10:54:41.675474 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mcjzc" podStartSLOduration=82.8890943 podStartE2EDuration="1m26.675448157s" podCreationTimestamp="2025-12-05 10:53:15 +0000 UTC" firstStartedPulling="2025-12-05 10:54:37.578671333 +0000 UTC m=+437.581865184" lastFinishedPulling="2025-12-05 10:54:41.36502518 +0000 UTC m=+441.368219041" observedRunningTime="2025-12-05 10:54:41.669964571 +0000 UTC m=+441.673158432" watchObservedRunningTime="2025-12-05 10:54:41.675448157 +0000 UTC m=+441.678642018" Dec 05 10:54:41.687886 master-0 kubenswrapper[24928]: I1205 10:54:41.687832 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52wjg" event={"ID":"241003b6-4aa6-4ed5-8bdd-46f2d30662fd","Type":"ContainerStarted","Data":"6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c"} Dec 05 10:54:41.694376 master-0 kubenswrapper[24928]: I1205 10:54:41.694140 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l4grl" podStartSLOduration=84.874555575 podStartE2EDuration="1m28.693979097s" podCreationTimestamp="2025-12-05 10:53:13 +0000 UTC" firstStartedPulling="2025-12-05 10:54:37.544112093 +0000 UTC m=+437.547305944" lastFinishedPulling="2025-12-05 10:54:41.363535595 +0000 UTC m=+441.366729466" observedRunningTime="2025-12-05 10:54:41.692152694 +0000 UTC m=+441.695346545" watchObservedRunningTime="2025-12-05 10:54:41.693979097 +0000 UTC m=+441.697172968" Dec 05 10:54:41.706654 master-0 kubenswrapper[24928]: I1205 10:54:41.706601 24928 scope.go:117] "RemoveContainer" containerID="b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38" Dec 05 10:54:41.707128 master-0 kubenswrapper[24928]: E1205 10:54:41.707028 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38\": container with ID starting with b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38 not found: ID does not exist" containerID="b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38" Dec 05 10:54:41.707128 master-0 kubenswrapper[24928]: I1205 10:54:41.707064 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38"} err="failed to get container status \"b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38\": rpc error: code = NotFound desc = could not find container \"b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38\": container with ID starting with b0408c26be12565c2f2ea566bd4fb549104ac8c86f1b079c3d2d7dd0d6307c38 not found: ID does not exist" Dec 05 10:54:41.736518 master-0 kubenswrapper[24928]: I1205 10:54:41.736244 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8pb58" podStartSLOduration=84.946733716 podStartE2EDuration="1m28.736228725s" podCreationTimestamp="2025-12-05 10:53:13 +0000 UTC" firstStartedPulling="2025-12-05 10:54:37.575007828 +0000 UTC m=+437.578201679" lastFinishedPulling="2025-12-05 10:54:41.364502837 +0000 UTC m=+441.367696688" observedRunningTime="2025-12-05 10:54:41.733975273 +0000 UTC m=+441.737169124" watchObservedRunningTime="2025-12-05 10:54:41.736228725 +0000 UTC m=+441.739422576" Dec 05 10:54:42.221801 master-0 kubenswrapper[24928]: I1205 10:54:42.221750 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b83ccd6fa217a93a2c607d0109896ef8" path="/var/lib/kubelet/pods/b83ccd6fa217a93a2c607d0109896ef8/volumes" Dec 05 10:54:42.222066 master-0 kubenswrapper[24928]: I1205 10:54:42.222041 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" podUID="" Dec 05 10:54:42.248483 master-0 kubenswrapper[24928]: I1205 10:54:42.248416 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0"] Dec 05 10:54:42.248576 master-0 kubenswrapper[24928]: I1205 10:54:42.248481 24928 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" mirrorPodUID="8d6043cc-e128-42ab-9b43-621c6ce10510" Dec 05 10:54:42.254711 master-0 kubenswrapper[24928]: I1205 10:54:42.254605 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0"] Dec 05 10:54:42.254711 master-0 kubenswrapper[24928]: I1205 10:54:42.254653 24928 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-master-0" mirrorPodUID="8d6043cc-e128-42ab-9b43-621c6ce10510" Dec 05 10:54:42.717447 master-0 kubenswrapper[24928]: I1205 10:54:42.715610 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" event={"ID":"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7","Type":"ContainerStarted","Data":"9fd91b1150a0d90199e526f726e8ef430696a6b00959a29a8921c711187b93f2"} Dec 05 10:54:42.717447 master-0 kubenswrapper[24928]: I1205 10:54:42.715661 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" event={"ID":"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7","Type":"ContainerStarted","Data":"14a8464711c65b375d4309f3c575404afec1966628e6d377b585d80e82912e32"} Dec 05 10:54:42.718773 master-0 kubenswrapper[24928]: I1205 10:54:42.718694 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"880060c7-79ed-428b-be59-ff663617615b","Type":"ContainerStarted","Data":"7354df277959541662cb9fb158b111fc70770db6b808abeafdd8996e3bdbc1a9"} Dec 05 10:54:42.718832 master-0 kubenswrapper[24928]: I1205 10:54:42.718785 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"880060c7-79ed-428b-be59-ff663617615b","Type":"ContainerStarted","Data":"7735bc5515748b77793639779a00ccedb431dc376e19ceff401f54873ea6399d"} Dec 05 10:54:42.718832 master-0 kubenswrapper[24928]: I1205 10:54:42.718800 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"880060c7-79ed-428b-be59-ff663617615b","Type":"ContainerStarted","Data":"8989c0560e8ff72d1c4b37517122ed6b04d351fb937daed052ac6668df7397d4"} Dec 05 10:54:42.718832 master-0 kubenswrapper[24928]: I1205 10:54:42.718810 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"880060c7-79ed-428b-be59-ff663617615b","Type":"ContainerStarted","Data":"0186e9a30b02b51d02622dc7a7f476d73368da579be2e2903ff99850f3879a73"} Dec 05 10:54:42.718832 master-0 kubenswrapper[24928]: I1205 10:54:42.718819 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"880060c7-79ed-428b-be59-ff663617615b","Type":"ContainerStarted","Data":"637b29b41dc787b460d7bd9c2c5ac2453e87b908d49a8570a714813300574ae5"} Dec 05 10:54:42.720089 master-0 kubenswrapper[24928]: I1205 10:54:42.720044 24928 generic.go:334] "Generic (PLEG): container finished" podID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" containerID="6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c" exitCode=0 Dec 05 10:54:42.720089 master-0 kubenswrapper[24928]: I1205 10:54:42.720092 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52wjg" event={"ID":"241003b6-4aa6-4ed5-8bdd-46f2d30662fd","Type":"ContainerDied","Data":"6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c"} Dec 05 10:54:42.729038 master-0 kubenswrapper[24928]: I1205 10:54:42.725412 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8pb58" event={"ID":"8a54075d-4cba-439e-96c0-db16e991d658","Type":"ContainerStarted","Data":"af68b8278694217e9435580393ef2485614c877c89eaa1ccd5cbe21efd7c6a14"} Dec 05 10:54:42.729038 master-0 kubenswrapper[24928]: I1205 10:54:42.728684 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-kfn9l" Dec 05 10:54:42.732802 master-0 kubenswrapper[24928]: I1205 10:54:42.730999 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"97818d19-9a0b-40a3-a075-a3ed60eebe62","Type":"ContainerStarted","Data":"8754adb16a8be65b606e170d307cfe47bf81c540a04c28fc298ff89db296ec94"} Dec 05 10:54:42.732802 master-0 kubenswrapper[24928]: I1205 10:54:42.731046 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"97818d19-9a0b-40a3-a075-a3ed60eebe62","Type":"ContainerStarted","Data":"5d116c328a1f4226c388840faa64310f72fc0238deecb61739137968600cc7ca"} Dec 05 10:54:42.732802 master-0 kubenswrapper[24928]: I1205 10:54:42.731061 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"97818d19-9a0b-40a3-a075-a3ed60eebe62","Type":"ContainerStarted","Data":"80466dd0b3d6e76339f4f0c64bb58bb74f05ed95595585f1e074a64dddd0d07e"} Dec 05 10:54:42.732802 master-0 kubenswrapper[24928]: I1205 10:54:42.731073 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"97818d19-9a0b-40a3-a075-a3ed60eebe62","Type":"ContainerStarted","Data":"f46eedf63897c70075f6b4dc97b21a4ad918312bb60701e6aca44a8474acd098"} Dec 05 10:54:42.732802 master-0 kubenswrapper[24928]: I1205 10:54:42.731085 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"97818d19-9a0b-40a3-a075-a3ed60eebe62","Type":"ContainerStarted","Data":"0fb1ba16c35123baf22bba4f16fc9e4c9ab7c9303efbb077be41eb0a7188651d"} Dec 05 10:54:42.732802 master-0 kubenswrapper[24928]: I1205 10:54:42.731096 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"97818d19-9a0b-40a3-a075-a3ed60eebe62","Type":"ContainerStarted","Data":"3ac838d85c99b4991f089991305cfe356fd7aa56734b246345561fda218c3ad7"} Dec 05 10:54:42.782513 master-0 kubenswrapper[24928]: I1205 10:54:42.778771 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=63.938400873 podStartE2EDuration="1m8.778723155s" podCreationTimestamp="2025-12-05 10:53:34 +0000 UTC" firstStartedPulling="2025-12-05 10:54:36.520568402 +0000 UTC m=+436.523762253" lastFinishedPulling="2025-12-05 10:54:41.360890684 +0000 UTC m=+441.364084535" observedRunningTime="2025-12-05 10:54:42.77548025 +0000 UTC m=+442.778674111" watchObservedRunningTime="2025-12-05 10:54:42.778723155 +0000 UTC m=+442.781917006" Dec 05 10:54:43.745535 master-0 kubenswrapper[24928]: I1205 10:54:43.745388 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" event={"ID":"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7","Type":"ContainerStarted","Data":"1ea7e04ebbf29f2b7647e28573e8410f6d4482bc97eeb01276e7b396b8e0cac4"} Dec 05 10:54:43.745535 master-0 kubenswrapper[24928]: I1205 10:54:43.745473 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" event={"ID":"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7","Type":"ContainerStarted","Data":"a2b8c944289c922eb85c7bc1531ee1781346a7a807ff7979fda8b27f29e34053"} Dec 05 10:54:43.745535 master-0 kubenswrapper[24928]: I1205 10:54:43.745488 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" event={"ID":"c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7","Type":"ContainerStarted","Data":"b6518dc0bf702df22a3d35c08980dd71006a07dd9c74ac370b1ee90ffaab461d"} Dec 05 10:54:43.746702 master-0 kubenswrapper[24928]: I1205 10:54:43.745615 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:43.749496 master-0 kubenswrapper[24928]: I1205 10:54:43.749414 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"880060c7-79ed-428b-be59-ff663617615b","Type":"ContainerStarted","Data":"e1669f38ed5fdc9878d1f7c1dd756d6aaa453c9fcd93ca47ad3ce842734b63e0"} Dec 05 10:54:43.752914 master-0 kubenswrapper[24928]: I1205 10:54:43.752868 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52wjg" event={"ID":"241003b6-4aa6-4ed5-8bdd-46f2d30662fd","Type":"ContainerStarted","Data":"78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd"} Dec 05 10:54:43.777252 master-0 kubenswrapper[24928]: I1205 10:54:43.777159 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" podStartSLOduration=63.24425303 podStartE2EDuration="1m9.777133905s" podCreationTimestamp="2025-12-05 10:53:34 +0000 UTC" firstStartedPulling="2025-12-05 10:54:36.431296845 +0000 UTC m=+436.434490696" lastFinishedPulling="2025-12-05 10:54:42.96417773 +0000 UTC m=+442.967371571" observedRunningTime="2025-12-05 10:54:43.771557706 +0000 UTC m=+443.774751567" watchObservedRunningTime="2025-12-05 10:54:43.777133905 +0000 UTC m=+443.780327776" Dec 05 10:54:43.811461 master-0 kubenswrapper[24928]: I1205 10:54:43.811262 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=64.039439753 podStartE2EDuration="1m9.811242594s" podCreationTimestamp="2025-12-05 10:53:34 +0000 UTC" firstStartedPulling="2025-12-05 10:54:37.558268951 +0000 UTC m=+437.561462802" lastFinishedPulling="2025-12-05 10:54:43.330071792 +0000 UTC m=+443.333265643" observedRunningTime="2025-12-05 10:54:43.809294169 +0000 UTC m=+443.812488020" watchObservedRunningTime="2025-12-05 10:54:43.811242594 +0000 UTC m=+443.814436445" Dec 05 10:54:43.842405 master-0 kubenswrapper[24928]: I1205 10:54:43.842267 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-52wjg" podStartSLOduration=81.911157062 podStartE2EDuration="1m27.842228001s" podCreationTimestamp="2025-12-05 10:53:16 +0000 UTC" firstStartedPulling="2025-12-05 10:54:37.549623651 +0000 UTC m=+437.552817502" lastFinishedPulling="2025-12-05 10:54:43.48069458 +0000 UTC m=+443.483888441" observedRunningTime="2025-12-05 10:54:43.84000666 +0000 UTC m=+443.843200521" watchObservedRunningTime="2025-12-05 10:54:43.842228001 +0000 UTC m=+443.845421852" Dec 05 10:54:45.187847 master-0 kubenswrapper[24928]: I1205 10:54:45.187760 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:54:45.253221 master-0 kubenswrapper[24928]: I1205 10:54:45.252820 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:45.253221 master-0 kubenswrapper[24928]: I1205 10:54:45.252915 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:45.259338 master-0 kubenswrapper[24928]: I1205 10:54:45.259236 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:45.390951 master-0 kubenswrapper[24928]: I1205 10:54:45.390839 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:45.390951 master-0 kubenswrapper[24928]: I1205 10:54:45.390904 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:45.417823 master-0 kubenswrapper[24928]: I1205 10:54:45.417749 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:45.417823 master-0 kubenswrapper[24928]: I1205 10:54:45.417809 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:45.433604 master-0 kubenswrapper[24928]: I1205 10:54:45.433518 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:45.433604 master-0 kubenswrapper[24928]: I1205 10:54:45.433598 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:45.436385 master-0 kubenswrapper[24928]: I1205 10:54:45.436321 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:45.469509 master-0 kubenswrapper[24928]: I1205 10:54:45.467811 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:45.469509 master-0 kubenswrapper[24928]: I1205 10:54:45.467919 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:45.514270 master-0 kubenswrapper[24928]: I1205 10:54:45.514214 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:45.523103 master-0 kubenswrapper[24928]: I1205 10:54:45.523054 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:45.600215 master-0 kubenswrapper[24928]: I1205 10:54:45.599980 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:45.600215 master-0 kubenswrapper[24928]: I1205 10:54:45.600038 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:45.605798 master-0 kubenswrapper[24928]: I1205 10:54:45.605750 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:45.772208 master-0 kubenswrapper[24928]: I1205 10:54:45.772084 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:54:45.772715 master-0 kubenswrapper[24928]: I1205 10:54:45.772673 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:54:46.463542 master-0 kubenswrapper[24928]: I1205 10:54:46.463408 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8pb58" podUID="8a54075d-4cba-439e-96c0-db16e991d658" containerName="registry-server" probeResult="failure" output=< Dec 05 10:54:46.463542 master-0 kubenswrapper[24928]: timeout: failed to connect service ":50051" within 1s Dec 05 10:54:46.463542 master-0 kubenswrapper[24928]: > Dec 05 10:54:46.847913 master-0 kubenswrapper[24928]: I1205 10:54:46.847845 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:54:47.695821 master-0 kubenswrapper[24928]: I1205 10:54:47.695743 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-74977ddd8b-dkrkh"] Dec 05 10:54:50.360898 master-0 kubenswrapper[24928]: I1205 10:54:50.360839 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-598896584f-9pd95" Dec 05 10:54:55.390455 master-0 kubenswrapper[24928]: I1205 10:54:55.390379 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:55.390455 master-0 kubenswrapper[24928]: I1205 10:54:55.390465 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:54:55.457037 master-0 kubenswrapper[24928]: I1205 10:54:55.456981 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:55.483747 master-0 kubenswrapper[24928]: I1205 10:54:55.483689 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:54:55.496698 master-0 kubenswrapper[24928]: I1205 10:54:55.496653 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:54:55.511270 master-0 kubenswrapper[24928]: I1205 10:54:55.511200 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:54:59.854847 master-0 kubenswrapper[24928]: I1205 10:54:59.854701 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-79cdddb8b4-mwjwx" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" containerID="cri-o://cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320" gracePeriod=15 Dec 05 10:55:00.377760 master-0 kubenswrapper[24928]: I1205 10:55:00.377673 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-79cdddb8b4-mwjwx_ac38876d-122b-4146-babf-722930c2c4ae/console/1.log" Dec 05 10:55:00.378505 master-0 kubenswrapper[24928]: I1205 10:55:00.378409 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-79cdddb8b4-mwjwx_ac38876d-122b-4146-babf-722930c2c4ae/console/0.log" Dec 05 10:55:00.378656 master-0 kubenswrapper[24928]: I1205 10:55:00.378570 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:55:00.433217 master-0 kubenswrapper[24928]: I1205 10:55:00.433137 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-console-config\") pod \"ac38876d-122b-4146-babf-722930c2c4ae\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " Dec 05 10:55:00.433217 master-0 kubenswrapper[24928]: I1205 10:55:00.433200 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-service-ca\") pod \"ac38876d-122b-4146-babf-722930c2c4ae\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " Dec 05 10:55:00.433544 master-0 kubenswrapper[24928]: I1205 10:55:00.433242 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vmlc\" (UniqueName: \"kubernetes.io/projected/ac38876d-122b-4146-babf-722930c2c4ae-kube-api-access-7vmlc\") pod \"ac38876d-122b-4146-babf-722930c2c4ae\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " Dec 05 10:55:00.433544 master-0 kubenswrapper[24928]: I1205 10:55:00.433336 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-serving-cert\") pod \"ac38876d-122b-4146-babf-722930c2c4ae\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " Dec 05 10:55:00.433544 master-0 kubenswrapper[24928]: I1205 10:55:00.433447 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-oauth-config\") pod \"ac38876d-122b-4146-babf-722930c2c4ae\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " Dec 05 10:55:00.433544 master-0 kubenswrapper[24928]: I1205 10:55:00.433491 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-trusted-ca-bundle\") pod \"ac38876d-122b-4146-babf-722930c2c4ae\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " Dec 05 10:55:00.433544 master-0 kubenswrapper[24928]: I1205 10:55:00.433517 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-oauth-serving-cert\") pod \"ac38876d-122b-4146-babf-722930c2c4ae\" (UID: \"ac38876d-122b-4146-babf-722930c2c4ae\") " Dec 05 10:55:00.434215 master-0 kubenswrapper[24928]: I1205 10:55:00.434135 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-console-config" (OuterVolumeSpecName: "console-config") pod "ac38876d-122b-4146-babf-722930c2c4ae" (UID: "ac38876d-122b-4146-babf-722930c2c4ae"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:55:00.434384 master-0 kubenswrapper[24928]: I1205 10:55:00.434339 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "ac38876d-122b-4146-babf-722930c2c4ae" (UID: "ac38876d-122b-4146-babf-722930c2c4ae"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:55:00.434384 master-0 kubenswrapper[24928]: I1205 10:55:00.434358 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-service-ca" (OuterVolumeSpecName: "service-ca") pod "ac38876d-122b-4146-babf-722930c2c4ae" (UID: "ac38876d-122b-4146-babf-722930c2c4ae"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:55:00.434580 master-0 kubenswrapper[24928]: I1205 10:55:00.434400 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "ac38876d-122b-4146-babf-722930c2c4ae" (UID: "ac38876d-122b-4146-babf-722930c2c4ae"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:55:00.436702 master-0 kubenswrapper[24928]: I1205 10:55:00.436584 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "ac38876d-122b-4146-babf-722930c2c4ae" (UID: "ac38876d-122b-4146-babf-722930c2c4ae"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:55:00.437792 master-0 kubenswrapper[24928]: I1205 10:55:00.437699 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac38876d-122b-4146-babf-722930c2c4ae-kube-api-access-7vmlc" (OuterVolumeSpecName: "kube-api-access-7vmlc") pod "ac38876d-122b-4146-babf-722930c2c4ae" (UID: "ac38876d-122b-4146-babf-722930c2c4ae"). InnerVolumeSpecName "kube-api-access-7vmlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:55:00.438699 master-0 kubenswrapper[24928]: I1205 10:55:00.438665 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "ac38876d-122b-4146-babf-722930c2c4ae" (UID: "ac38876d-122b-4146-babf-722930c2c4ae"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:55:00.535623 master-0 kubenswrapper[24928]: I1205 10:55:00.535558 24928 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:00.535952 master-0 kubenswrapper[24928]: I1205 10:55:00.535934 24928 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ac38876d-122b-4146-babf-722930c2c4ae-console-oauth-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:00.536057 master-0 kubenswrapper[24928]: I1205 10:55:00.536040 24928 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-trusted-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:00.536153 master-0 kubenswrapper[24928]: I1205 10:55:00.536136 24928 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-oauth-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:00.536249 master-0 kubenswrapper[24928]: I1205 10:55:00.536233 24928 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-console-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:00.536353 master-0 kubenswrapper[24928]: I1205 10:55:00.536337 24928 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ac38876d-122b-4146-babf-722930c2c4ae-service-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:00.536562 master-0 kubenswrapper[24928]: I1205 10:55:00.536543 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vmlc\" (UniqueName: \"kubernetes.io/projected/ac38876d-122b-4146-babf-722930c2c4ae-kube-api-access-7vmlc\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:00.909600 master-0 kubenswrapper[24928]: I1205 10:55:00.909486 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-79cdddb8b4-mwjwx_ac38876d-122b-4146-babf-722930c2c4ae/console/1.log" Dec 05 10:55:00.911031 master-0 kubenswrapper[24928]: I1205 10:55:00.910957 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-79cdddb8b4-mwjwx_ac38876d-122b-4146-babf-722930c2c4ae/console/0.log" Dec 05 10:55:00.911269 master-0 kubenswrapper[24928]: I1205 10:55:00.911037 24928 generic.go:334] "Generic (PLEG): container finished" podID="ac38876d-122b-4146-babf-722930c2c4ae" containerID="cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320" exitCode=2 Dec 05 10:55:00.911269 master-0 kubenswrapper[24928]: I1205 10:55:00.911070 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79cdddb8b4-mwjwx" event={"ID":"ac38876d-122b-4146-babf-722930c2c4ae","Type":"ContainerDied","Data":"cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320"} Dec 05 10:55:00.911269 master-0 kubenswrapper[24928]: I1205 10:55:00.911131 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-79cdddb8b4-mwjwx" event={"ID":"ac38876d-122b-4146-babf-722930c2c4ae","Type":"ContainerDied","Data":"232db71c9974b1ab9b26647298ec9c771d701ecd59647a4c717df6ca62c916c3"} Dec 05 10:55:00.911269 master-0 kubenswrapper[24928]: I1205 10:55:00.911146 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-79cdddb8b4-mwjwx" Dec 05 10:55:00.911763 master-0 kubenswrapper[24928]: I1205 10:55:00.911153 24928 scope.go:117] "RemoveContainer" containerID="cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320" Dec 05 10:55:00.953541 master-0 kubenswrapper[24928]: I1205 10:55:00.953345 24928 scope.go:117] "RemoveContainer" containerID="4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02" Dec 05 10:55:00.978519 master-0 kubenswrapper[24928]: I1205 10:55:00.972174 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-79cdddb8b4-mwjwx"] Dec 05 10:55:00.978519 master-0 kubenswrapper[24928]: I1205 10:55:00.977530 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-79cdddb8b4-mwjwx"] Dec 05 10:55:00.987687 master-0 kubenswrapper[24928]: I1205 10:55:00.987648 24928 scope.go:117] "RemoveContainer" containerID="cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320" Dec 05 10:55:00.988088 master-0 kubenswrapper[24928]: E1205 10:55:00.988037 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320\": container with ID starting with cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320 not found: ID does not exist" containerID="cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320" Dec 05 10:55:00.988169 master-0 kubenswrapper[24928]: I1205 10:55:00.988090 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320"} err="failed to get container status \"cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320\": rpc error: code = NotFound desc = could not find container \"cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320\": container with ID starting with cb7f4192fc01e5a67e2d232f37b71499793f74f9e5a5e91cee3abe3541382320 not found: ID does not exist" Dec 05 10:55:00.988169 master-0 kubenswrapper[24928]: I1205 10:55:00.988121 24928 scope.go:117] "RemoveContainer" containerID="4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02" Dec 05 10:55:00.988405 master-0 kubenswrapper[24928]: E1205 10:55:00.988379 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02\": container with ID starting with 4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02 not found: ID does not exist" containerID="4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02" Dec 05 10:55:00.988500 master-0 kubenswrapper[24928]: I1205 10:55:00.988404 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02"} err="failed to get container status \"4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02\": rpc error: code = NotFound desc = could not find container \"4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02\": container with ID starting with 4ca806d3f4af3aa5c7a4362c6b85394f5f675051f071f12e1055e0724c169b02 not found: ID does not exist" Dec 05 10:55:01.204743 master-0 kubenswrapper[24928]: I1205 10:55:01.204611 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:55:01.204743 master-0 kubenswrapper[24928]: I1205 10:55:01.204673 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:55:01.230302 master-0 kubenswrapper[24928]: I1205 10:55:01.230235 24928 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-etcd/etcd-master-0" Dec 05 10:55:01.240973 master-0 kubenswrapper[24928]: I1205 10:55:01.240933 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-etcd/etcd-master-0"] Dec 05 10:55:01.252355 master-0 kubenswrapper[24928]: I1205 10:55:01.252263 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-etcd/etcd-master-0"] Dec 05 10:55:01.270267 master-0 kubenswrapper[24928]: I1205 10:55:01.270195 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-master-0"] Dec 05 10:55:01.922789 master-0 kubenswrapper[24928]: I1205 10:55:01.922687 24928 kubelet.go:1909] "Trying to delete pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:55:01.922789 master-0 kubenswrapper[24928]: I1205 10:55:01.922737 24928 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-etcd/etcd-master-0" podUID="67149a2d-9332-4d46-91d4-bd0bef169d9b" Dec 05 10:55:02.213280 master-0 kubenswrapper[24928]: I1205 10:55:02.213040 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac38876d-122b-4146-babf-722930c2c4ae" path="/var/lib/kubelet/pods/ac38876d-122b-4146-babf-722930c2c4ae/volumes" Dec 05 10:55:07.273456 master-0 kubenswrapper[24928]: I1205 10:55:07.273352 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-np6r8"] Dec 05 10:55:07.274070 master-0 kubenswrapper[24928]: E1205 10:55:07.273716 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b83ccd6fa217a93a2c607d0109896ef8" containerName="startup-monitor" Dec 05 10:55:07.274070 master-0 kubenswrapper[24928]: I1205 10:55:07.273738 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b83ccd6fa217a93a2c607d0109896ef8" containerName="startup-monitor" Dec 05 10:55:07.274070 master-0 kubenswrapper[24928]: E1205 10:55:07.273770 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" Dec 05 10:55:07.274070 master-0 kubenswrapper[24928]: I1205 10:55:07.273784 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" Dec 05 10:55:07.274070 master-0 kubenswrapper[24928]: E1205 10:55:07.273818 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" Dec 05 10:55:07.274070 master-0 kubenswrapper[24928]: I1205 10:55:07.273830 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" Dec 05 10:55:07.274070 master-0 kubenswrapper[24928]: I1205 10:55:07.274057 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" Dec 05 10:55:07.274343 master-0 kubenswrapper[24928]: I1205 10:55:07.274094 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac38876d-122b-4146-babf-722930c2c4ae" containerName="console" Dec 05 10:55:07.274343 master-0 kubenswrapper[24928]: I1205 10:55:07.274155 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="b83ccd6fa217a93a2c607d0109896ef8" containerName="startup-monitor" Dec 05 10:55:07.274779 master-0 kubenswrapper[24928]: I1205 10:55:07.274745 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-np6r8" Dec 05 10:55:07.277037 master-0 kubenswrapper[24928]: I1205 10:55:07.276997 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 05 10:55:07.315584 master-0 kubenswrapper[24928]: I1205 10:55:07.315475 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-master-0" podStartSLOduration=6.3154025560000004 podStartE2EDuration="6.315402556s" podCreationTimestamp="2025-12-05 10:55:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:55:07.311026175 +0000 UTC m=+467.314220026" watchObservedRunningTime="2025-12-05 10:55:07.315402556 +0000 UTC m=+467.318596407" Dec 05 10:55:07.353765 master-0 kubenswrapper[24928]: I1205 10:55:07.353679 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd2ns\" (UniqueName: \"kubernetes.io/projected/dbafe605-f87b-47c6-9aa4-ae996044461c-kube-api-access-gd2ns\") pod \"node-ca-np6r8\" (UID: \"dbafe605-f87b-47c6-9aa4-ae996044461c\") " pod="openshift-image-registry/node-ca-np6r8" Dec 05 10:55:07.353765 master-0 kubenswrapper[24928]: I1205 10:55:07.353752 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dbafe605-f87b-47c6-9aa4-ae996044461c-host\") pod \"node-ca-np6r8\" (UID: \"dbafe605-f87b-47c6-9aa4-ae996044461c\") " pod="openshift-image-registry/node-ca-np6r8" Dec 05 10:55:07.354003 master-0 kubenswrapper[24928]: I1205 10:55:07.353934 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/dbafe605-f87b-47c6-9aa4-ae996044461c-serviceca\") pod \"node-ca-np6r8\" (UID: \"dbafe605-f87b-47c6-9aa4-ae996044461c\") " pod="openshift-image-registry/node-ca-np6r8" Dec 05 10:55:07.454967 master-0 kubenswrapper[24928]: I1205 10:55:07.454867 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/dbafe605-f87b-47c6-9aa4-ae996044461c-serviceca\") pod \"node-ca-np6r8\" (UID: \"dbafe605-f87b-47c6-9aa4-ae996044461c\") " pod="openshift-image-registry/node-ca-np6r8" Dec 05 10:55:07.455289 master-0 kubenswrapper[24928]: I1205 10:55:07.455085 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd2ns\" (UniqueName: \"kubernetes.io/projected/dbafe605-f87b-47c6-9aa4-ae996044461c-kube-api-access-gd2ns\") pod \"node-ca-np6r8\" (UID: \"dbafe605-f87b-47c6-9aa4-ae996044461c\") " pod="openshift-image-registry/node-ca-np6r8" Dec 05 10:55:07.455519 master-0 kubenswrapper[24928]: I1205 10:55:07.455376 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dbafe605-f87b-47c6-9aa4-ae996044461c-host\") pod \"node-ca-np6r8\" (UID: \"dbafe605-f87b-47c6-9aa4-ae996044461c\") " pod="openshift-image-registry/node-ca-np6r8" Dec 05 10:55:07.455692 master-0 kubenswrapper[24928]: I1205 10:55:07.455415 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/dbafe605-f87b-47c6-9aa4-ae996044461c-host\") pod \"node-ca-np6r8\" (UID: \"dbafe605-f87b-47c6-9aa4-ae996044461c\") " pod="openshift-image-registry/node-ca-np6r8" Dec 05 10:55:07.456167 master-0 kubenswrapper[24928]: I1205 10:55:07.456113 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/dbafe605-f87b-47c6-9aa4-ae996044461c-serviceca\") pod \"node-ca-np6r8\" (UID: \"dbafe605-f87b-47c6-9aa4-ae996044461c\") " pod="openshift-image-registry/node-ca-np6r8" Dec 05 10:55:07.471157 master-0 kubenswrapper[24928]: I1205 10:55:07.471083 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd2ns\" (UniqueName: \"kubernetes.io/projected/dbafe605-f87b-47c6-9aa4-ae996044461c-kube-api-access-gd2ns\") pod \"node-ca-np6r8\" (UID: \"dbafe605-f87b-47c6-9aa4-ae996044461c\") " pod="openshift-image-registry/node-ca-np6r8" Dec 05 10:55:07.597467 master-0 kubenswrapper[24928]: I1205 10:55:07.597343 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-np6r8" Dec 05 10:55:07.644340 master-0 kubenswrapper[24928]: W1205 10:55:07.644252 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbafe605_f87b_47c6_9aa4_ae996044461c.slice/crio-82f062cdb39d693ddd99f77b83cb7fc57d66e1953ff550b3dd5504a758be0663 WatchSource:0}: Error finding container 82f062cdb39d693ddd99f77b83cb7fc57d66e1953ff550b3dd5504a758be0663: Status 404 returned error can't find the container with id 82f062cdb39d693ddd99f77b83cb7fc57d66e1953ff550b3dd5504a758be0663 Dec 05 10:55:07.979122 master-0 kubenswrapper[24928]: I1205 10:55:07.978910 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-np6r8" event={"ID":"dbafe605-f87b-47c6-9aa4-ae996044461c","Type":"ContainerStarted","Data":"82f062cdb39d693ddd99f77b83cb7fc57d66e1953ff550b3dd5504a758be0663"} Dec 05 10:55:09.990368 master-0 kubenswrapper[24928]: I1205 10:55:09.990301 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-52wjg"] Dec 05 10:55:09.991784 master-0 kubenswrapper[24928]: I1205 10:55:09.990622 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-52wjg" podUID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" containerName="registry-server" containerID="cri-o://78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd" gracePeriod=2 Dec 05 10:55:10.188701 master-0 kubenswrapper[24928]: I1205 10:55:10.188629 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mcjzc"] Dec 05 10:55:10.188941 master-0 kubenswrapper[24928]: I1205 10:55:10.188898 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mcjzc" podUID="25d480c2-7108-450d-adc7-c56e0b5be0c8" containerName="registry-server" containerID="cri-o://1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2" gracePeriod=2 Dec 05 10:55:10.508724 master-0 kubenswrapper[24928]: I1205 10:55:10.508665 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:55:10.626842 master-0 kubenswrapper[24928]: I1205 10:55:10.626731 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-utilities\") pod \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " Dec 05 10:55:10.626842 master-0 kubenswrapper[24928]: I1205 10:55:10.626806 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-catalog-content\") pod \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " Dec 05 10:55:10.627066 master-0 kubenswrapper[24928]: I1205 10:55:10.626982 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6474b\" (UniqueName: \"kubernetes.io/projected/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-kube-api-access-6474b\") pod \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\" (UID: \"241003b6-4aa6-4ed5-8bdd-46f2d30662fd\") " Dec 05 10:55:10.627602 master-0 kubenswrapper[24928]: I1205 10:55:10.627572 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-utilities" (OuterVolumeSpecName: "utilities") pod "241003b6-4aa6-4ed5-8bdd-46f2d30662fd" (UID: "241003b6-4aa6-4ed5-8bdd-46f2d30662fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:55:10.630934 master-0 kubenswrapper[24928]: I1205 10:55:10.630863 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-kube-api-access-6474b" (OuterVolumeSpecName: "kube-api-access-6474b") pod "241003b6-4aa6-4ed5-8bdd-46f2d30662fd" (UID: "241003b6-4aa6-4ed5-8bdd-46f2d30662fd"). InnerVolumeSpecName "kube-api-access-6474b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:55:10.669207 master-0 kubenswrapper[24928]: I1205 10:55:10.669139 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "241003b6-4aa6-4ed5-8bdd-46f2d30662fd" (UID: "241003b6-4aa6-4ed5-8bdd-46f2d30662fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:55:10.702094 master-0 kubenswrapper[24928]: I1205 10:55:10.702049 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:55:10.729145 master-0 kubenswrapper[24928]: I1205 10:55:10.728875 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6474b\" (UniqueName: \"kubernetes.io/projected/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-kube-api-access-6474b\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:10.729145 master-0 kubenswrapper[24928]: I1205 10:55:10.728948 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:10.729145 master-0 kubenswrapper[24928]: I1205 10:55:10.728965 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/241003b6-4aa6-4ed5-8bdd-46f2d30662fd-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:10.831335 master-0 kubenswrapper[24928]: I1205 10:55:10.831223 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d54hx\" (UniqueName: \"kubernetes.io/projected/25d480c2-7108-450d-adc7-c56e0b5be0c8-kube-api-access-d54hx\") pod \"25d480c2-7108-450d-adc7-c56e0b5be0c8\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " Dec 05 10:55:10.831634 master-0 kubenswrapper[24928]: I1205 10:55:10.831568 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-utilities\") pod \"25d480c2-7108-450d-adc7-c56e0b5be0c8\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " Dec 05 10:55:10.831720 master-0 kubenswrapper[24928]: I1205 10:55:10.831684 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-catalog-content\") pod \"25d480c2-7108-450d-adc7-c56e0b5be0c8\" (UID: \"25d480c2-7108-450d-adc7-c56e0b5be0c8\") " Dec 05 10:55:10.832550 master-0 kubenswrapper[24928]: I1205 10:55:10.832497 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-utilities" (OuterVolumeSpecName: "utilities") pod "25d480c2-7108-450d-adc7-c56e0b5be0c8" (UID: "25d480c2-7108-450d-adc7-c56e0b5be0c8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:55:10.836052 master-0 kubenswrapper[24928]: I1205 10:55:10.835984 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25d480c2-7108-450d-adc7-c56e0b5be0c8-kube-api-access-d54hx" (OuterVolumeSpecName: "kube-api-access-d54hx") pod "25d480c2-7108-450d-adc7-c56e0b5be0c8" (UID: "25d480c2-7108-450d-adc7-c56e0b5be0c8"). InnerVolumeSpecName "kube-api-access-d54hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:55:10.877691 master-0 kubenswrapper[24928]: I1205 10:55:10.877575 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25d480c2-7108-450d-adc7-c56e0b5be0c8" (UID: "25d480c2-7108-450d-adc7-c56e0b5be0c8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:55:10.933271 master-0 kubenswrapper[24928]: I1205 10:55:10.933189 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:10.933271 master-0 kubenswrapper[24928]: I1205 10:55:10.933230 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d54hx\" (UniqueName: \"kubernetes.io/projected/25d480c2-7108-450d-adc7-c56e0b5be0c8-kube-api-access-d54hx\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:10.933271 master-0 kubenswrapper[24928]: I1205 10:55:10.933241 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25d480c2-7108-450d-adc7-c56e0b5be0c8-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:11.007588 master-0 kubenswrapper[24928]: I1205 10:55:11.007498 24928 generic.go:334] "Generic (PLEG): container finished" podID="25d480c2-7108-450d-adc7-c56e0b5be0c8" containerID="1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2" exitCode=0 Dec 05 10:55:11.007588 master-0 kubenswrapper[24928]: I1205 10:55:11.007569 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mcjzc" Dec 05 10:55:11.007588 master-0 kubenswrapper[24928]: I1205 10:55:11.007595 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcjzc" event={"ID":"25d480c2-7108-450d-adc7-c56e0b5be0c8","Type":"ContainerDied","Data":"1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2"} Dec 05 10:55:11.008319 master-0 kubenswrapper[24928]: I1205 10:55:11.007630 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mcjzc" event={"ID":"25d480c2-7108-450d-adc7-c56e0b5be0c8","Type":"ContainerDied","Data":"f87785d6016322c345dd513ef378b4a11bf62b7475beba1226df7371b9687bcc"} Dec 05 10:55:11.008319 master-0 kubenswrapper[24928]: I1205 10:55:11.007655 24928 scope.go:117] "RemoveContainer" containerID="1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2" Dec 05 10:55:11.010778 master-0 kubenswrapper[24928]: I1205 10:55:11.010730 24928 generic.go:334] "Generic (PLEG): container finished" podID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" containerID="78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd" exitCode=0 Dec 05 10:55:11.010869 master-0 kubenswrapper[24928]: I1205 10:55:11.010787 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52wjg" event={"ID":"241003b6-4aa6-4ed5-8bdd-46f2d30662fd","Type":"ContainerDied","Data":"78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd"} Dec 05 10:55:11.010869 master-0 kubenswrapper[24928]: I1205 10:55:11.010814 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-52wjg" event={"ID":"241003b6-4aa6-4ed5-8bdd-46f2d30662fd","Type":"ContainerDied","Data":"f5130e9cce93cfe214efea75ca82a1cbc656c3e820221aab863c357c84937203"} Dec 05 10:55:11.010964 master-0 kubenswrapper[24928]: I1205 10:55:11.010870 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-52wjg" Dec 05 10:55:11.012975 master-0 kubenswrapper[24928]: I1205 10:55:11.012922 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-np6r8" event={"ID":"dbafe605-f87b-47c6-9aa4-ae996044461c","Type":"ContainerStarted","Data":"c9b6a9cae3e8625ef42ad071d02c850a7340f30803ad933cddc60c1a83682846"} Dec 05 10:55:11.027169 master-0 kubenswrapper[24928]: I1205 10:55:11.026988 24928 scope.go:117] "RemoveContainer" containerID="da15035e6482378c271f4a44a0e9cb9b80044a94ce5194227563641492f09aa7" Dec 05 10:55:11.044458 master-0 kubenswrapper[24928]: I1205 10:55:11.044359 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-np6r8" podStartSLOduration=1.903778727 podStartE2EDuration="4.044340645s" podCreationTimestamp="2025-12-05 10:55:07 +0000 UTC" firstStartedPulling="2025-12-05 10:55:07.65456716 +0000 UTC m=+467.657761041" lastFinishedPulling="2025-12-05 10:55:09.795129108 +0000 UTC m=+469.798322959" observedRunningTime="2025-12-05 10:55:11.037785033 +0000 UTC m=+471.040978894" watchObservedRunningTime="2025-12-05 10:55:11.044340645 +0000 UTC m=+471.047534506" Dec 05 10:55:11.060173 master-0 kubenswrapper[24928]: I1205 10:55:11.060117 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mcjzc"] Dec 05 10:55:11.068796 master-0 kubenswrapper[24928]: I1205 10:55:11.068639 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mcjzc"] Dec 05 10:55:11.068908 master-0 kubenswrapper[24928]: I1205 10:55:11.068804 24928 scope.go:117] "RemoveContainer" containerID="f42999b81de2c99cb042a2f8cade529d3e0c6dc94e65a8e525d08f68560b9e73" Dec 05 10:55:11.074731 master-0 kubenswrapper[24928]: I1205 10:55:11.074638 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-52wjg"] Dec 05 10:55:11.090480 master-0 kubenswrapper[24928]: I1205 10:55:11.086001 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-52wjg"] Dec 05 10:55:11.117733 master-0 kubenswrapper[24928]: I1205 10:55:11.117664 24928 scope.go:117] "RemoveContainer" containerID="1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2" Dec 05 10:55:11.118221 master-0 kubenswrapper[24928]: E1205 10:55:11.118192 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2\": container with ID starting with 1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2 not found: ID does not exist" containerID="1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2" Dec 05 10:55:11.118273 master-0 kubenswrapper[24928]: I1205 10:55:11.118245 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2"} err="failed to get container status \"1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2\": rpc error: code = NotFound desc = could not find container \"1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2\": container with ID starting with 1cdcafca4abb31e11121ed7bf27a7cf67c1ec6fbc148eac7e9288d7deeb59ec2 not found: ID does not exist" Dec 05 10:55:11.118273 master-0 kubenswrapper[24928]: I1205 10:55:11.118267 24928 scope.go:117] "RemoveContainer" containerID="da15035e6482378c271f4a44a0e9cb9b80044a94ce5194227563641492f09aa7" Dec 05 10:55:11.118843 master-0 kubenswrapper[24928]: E1205 10:55:11.118815 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da15035e6482378c271f4a44a0e9cb9b80044a94ce5194227563641492f09aa7\": container with ID starting with da15035e6482378c271f4a44a0e9cb9b80044a94ce5194227563641492f09aa7 not found: ID does not exist" containerID="da15035e6482378c271f4a44a0e9cb9b80044a94ce5194227563641492f09aa7" Dec 05 10:55:11.118887 master-0 kubenswrapper[24928]: I1205 10:55:11.118841 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da15035e6482378c271f4a44a0e9cb9b80044a94ce5194227563641492f09aa7"} err="failed to get container status \"da15035e6482378c271f4a44a0e9cb9b80044a94ce5194227563641492f09aa7\": rpc error: code = NotFound desc = could not find container \"da15035e6482378c271f4a44a0e9cb9b80044a94ce5194227563641492f09aa7\": container with ID starting with da15035e6482378c271f4a44a0e9cb9b80044a94ce5194227563641492f09aa7 not found: ID does not exist" Dec 05 10:55:11.118887 master-0 kubenswrapper[24928]: I1205 10:55:11.118855 24928 scope.go:117] "RemoveContainer" containerID="f42999b81de2c99cb042a2f8cade529d3e0c6dc94e65a8e525d08f68560b9e73" Dec 05 10:55:11.119119 master-0 kubenswrapper[24928]: E1205 10:55:11.119090 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f42999b81de2c99cb042a2f8cade529d3e0c6dc94e65a8e525d08f68560b9e73\": container with ID starting with f42999b81de2c99cb042a2f8cade529d3e0c6dc94e65a8e525d08f68560b9e73 not found: ID does not exist" containerID="f42999b81de2c99cb042a2f8cade529d3e0c6dc94e65a8e525d08f68560b9e73" Dec 05 10:55:11.119119 master-0 kubenswrapper[24928]: I1205 10:55:11.119111 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f42999b81de2c99cb042a2f8cade529d3e0c6dc94e65a8e525d08f68560b9e73"} err="failed to get container status \"f42999b81de2c99cb042a2f8cade529d3e0c6dc94e65a8e525d08f68560b9e73\": rpc error: code = NotFound desc = could not find container \"f42999b81de2c99cb042a2f8cade529d3e0c6dc94e65a8e525d08f68560b9e73\": container with ID starting with f42999b81de2c99cb042a2f8cade529d3e0c6dc94e65a8e525d08f68560b9e73 not found: ID does not exist" Dec 05 10:55:11.119211 master-0 kubenswrapper[24928]: I1205 10:55:11.119124 24928 scope.go:117] "RemoveContainer" containerID="78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd" Dec 05 10:55:11.132927 master-0 kubenswrapper[24928]: I1205 10:55:11.132898 24928 scope.go:117] "RemoveContainer" containerID="6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c" Dec 05 10:55:11.147569 master-0 kubenswrapper[24928]: I1205 10:55:11.147517 24928 scope.go:117] "RemoveContainer" containerID="28abf4df43ff19287f182dc322a79951c1b7d10826c92fa80a92b4b4ba1bff28" Dec 05 10:55:11.172191 master-0 kubenswrapper[24928]: I1205 10:55:11.172129 24928 scope.go:117] "RemoveContainer" containerID="78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd" Dec 05 10:55:11.172858 master-0 kubenswrapper[24928]: E1205 10:55:11.172805 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd\": container with ID starting with 78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd not found: ID does not exist" containerID="78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd" Dec 05 10:55:11.172935 master-0 kubenswrapper[24928]: I1205 10:55:11.172857 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd"} err="failed to get container status \"78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd\": rpc error: code = NotFound desc = could not find container \"78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd\": container with ID starting with 78a336f6a81eb033de878314d920de4f7c48238e2581b7ef4802eba2bf5ba4bd not found: ID does not exist" Dec 05 10:55:11.172935 master-0 kubenswrapper[24928]: I1205 10:55:11.172891 24928 scope.go:117] "RemoveContainer" containerID="6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c" Dec 05 10:55:11.173445 master-0 kubenswrapper[24928]: E1205 10:55:11.173382 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c\": container with ID starting with 6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c not found: ID does not exist" containerID="6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c" Dec 05 10:55:11.173532 master-0 kubenswrapper[24928]: I1205 10:55:11.173415 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c"} err="failed to get container status \"6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c\": rpc error: code = NotFound desc = could not find container \"6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c\": container with ID starting with 6d76e152919a457726db62bf1d2bb75dea013b8dc862c029d078698fab98e01c not found: ID does not exist" Dec 05 10:55:11.173589 master-0 kubenswrapper[24928]: I1205 10:55:11.173534 24928 scope.go:117] "RemoveContainer" containerID="28abf4df43ff19287f182dc322a79951c1b7d10826c92fa80a92b4b4ba1bff28" Dec 05 10:55:11.173964 master-0 kubenswrapper[24928]: E1205 10:55:11.173902 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28abf4df43ff19287f182dc322a79951c1b7d10826c92fa80a92b4b4ba1bff28\": container with ID starting with 28abf4df43ff19287f182dc322a79951c1b7d10826c92fa80a92b4b4ba1bff28 not found: ID does not exist" containerID="28abf4df43ff19287f182dc322a79951c1b7d10826c92fa80a92b4b4ba1bff28" Dec 05 10:55:11.174032 master-0 kubenswrapper[24928]: I1205 10:55:11.173965 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28abf4df43ff19287f182dc322a79951c1b7d10826c92fa80a92b4b4ba1bff28"} err="failed to get container status \"28abf4df43ff19287f182dc322a79951c1b7d10826c92fa80a92b4b4ba1bff28\": rpc error: code = NotFound desc = could not find container \"28abf4df43ff19287f182dc322a79951c1b7d10826c92fa80a92b4b4ba1bff28\": container with ID starting with 28abf4df43ff19287f182dc322a79951c1b7d10826c92fa80a92b4b4ba1bff28 not found: ID does not exist" Dec 05 10:55:12.221238 master-0 kubenswrapper[24928]: I1205 10:55:12.221072 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" path="/var/lib/kubelet/pods/241003b6-4aa6-4ed5-8bdd-46f2d30662fd/volumes" Dec 05 10:55:12.222089 master-0 kubenswrapper[24928]: I1205 10:55:12.222037 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25d480c2-7108-450d-adc7-c56e0b5be0c8" path="/var/lib/kubelet/pods/25d480c2-7108-450d-adc7-c56e0b5be0c8/volumes" Dec 05 10:55:12.388767 master-0 kubenswrapper[24928]: I1205 10:55:12.388704 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4grl"] Dec 05 10:55:12.389038 master-0 kubenswrapper[24928]: I1205 10:55:12.388998 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l4grl" podUID="2934463a-d494-4735-a5d8-ad2c16aea065" containerName="registry-server" containerID="cri-o://2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9" gracePeriod=2 Dec 05 10:55:12.598834 master-0 kubenswrapper[24928]: I1205 10:55:12.598769 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8pb58"] Dec 05 10:55:12.599180 master-0 kubenswrapper[24928]: I1205 10:55:12.599069 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8pb58" podUID="8a54075d-4cba-439e-96c0-db16e991d658" containerName="registry-server" containerID="cri-o://af68b8278694217e9435580393ef2485614c877c89eaa1ccd5cbe21efd7c6a14" gracePeriod=2 Dec 05 10:55:12.792200 master-0 kubenswrapper[24928]: I1205 10:55:12.792157 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:55:12.835760 master-0 kubenswrapper[24928]: I1205 10:55:12.835671 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-74977ddd8b-dkrkh" podUID="629d199b-36e5-4c96-b58f-24314c4f2f58" containerName="console" containerID="cri-o://f1a948781cc195931bc78694a8e65f57a216949ec87c08ca993d58ff6b6fbbd1" gracePeriod=15 Dec 05 10:55:12.869991 master-0 kubenswrapper[24928]: I1205 10:55:12.869932 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-catalog-content\") pod \"2934463a-d494-4735-a5d8-ad2c16aea065\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " Dec 05 10:55:12.870127 master-0 kubenswrapper[24928]: I1205 10:55:12.870079 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xr7nj\" (UniqueName: \"kubernetes.io/projected/2934463a-d494-4735-a5d8-ad2c16aea065-kube-api-access-xr7nj\") pod \"2934463a-d494-4735-a5d8-ad2c16aea065\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " Dec 05 10:55:12.870127 master-0 kubenswrapper[24928]: I1205 10:55:12.870115 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-utilities\") pod \"2934463a-d494-4735-a5d8-ad2c16aea065\" (UID: \"2934463a-d494-4735-a5d8-ad2c16aea065\") " Dec 05 10:55:12.871528 master-0 kubenswrapper[24928]: I1205 10:55:12.871461 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-utilities" (OuterVolumeSpecName: "utilities") pod "2934463a-d494-4735-a5d8-ad2c16aea065" (UID: "2934463a-d494-4735-a5d8-ad2c16aea065"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:55:12.872978 master-0 kubenswrapper[24928]: I1205 10:55:12.872937 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2934463a-d494-4735-a5d8-ad2c16aea065-kube-api-access-xr7nj" (OuterVolumeSpecName: "kube-api-access-xr7nj") pod "2934463a-d494-4735-a5d8-ad2c16aea065" (UID: "2934463a-d494-4735-a5d8-ad2c16aea065"). InnerVolumeSpecName "kube-api-access-xr7nj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:55:12.889154 master-0 kubenswrapper[24928]: I1205 10:55:12.889090 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2934463a-d494-4735-a5d8-ad2c16aea065" (UID: "2934463a-d494-4735-a5d8-ad2c16aea065"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:55:12.972442 master-0 kubenswrapper[24928]: I1205 10:55:12.972213 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xr7nj\" (UniqueName: \"kubernetes.io/projected/2934463a-d494-4735-a5d8-ad2c16aea065-kube-api-access-xr7nj\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:12.972792 master-0 kubenswrapper[24928]: I1205 10:55:12.972750 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:12.972873 master-0 kubenswrapper[24928]: I1205 10:55:12.972794 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2934463a-d494-4735-a5d8-ad2c16aea065-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:13.071633 master-0 kubenswrapper[24928]: I1205 10:55:13.071503 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74977ddd8b-dkrkh_629d199b-36e5-4c96-b58f-24314c4f2f58/console/0.log" Dec 05 10:55:13.071633 master-0 kubenswrapper[24928]: I1205 10:55:13.071554 24928 generic.go:334] "Generic (PLEG): container finished" podID="629d199b-36e5-4c96-b58f-24314c4f2f58" containerID="f1a948781cc195931bc78694a8e65f57a216949ec87c08ca993d58ff6b6fbbd1" exitCode=2 Dec 05 10:55:13.071633 master-0 kubenswrapper[24928]: I1205 10:55:13.071600 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74977ddd8b-dkrkh" event={"ID":"629d199b-36e5-4c96-b58f-24314c4f2f58","Type":"ContainerDied","Data":"f1a948781cc195931bc78694a8e65f57a216949ec87c08ca993d58ff6b6fbbd1"} Dec 05 10:55:13.074084 master-0 kubenswrapper[24928]: I1205 10:55:13.074046 24928 generic.go:334] "Generic (PLEG): container finished" podID="8a54075d-4cba-439e-96c0-db16e991d658" containerID="af68b8278694217e9435580393ef2485614c877c89eaa1ccd5cbe21efd7c6a14" exitCode=0 Dec 05 10:55:13.074186 master-0 kubenswrapper[24928]: I1205 10:55:13.074104 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8pb58" event={"ID":"8a54075d-4cba-439e-96c0-db16e991d658","Type":"ContainerDied","Data":"af68b8278694217e9435580393ef2485614c877c89eaa1ccd5cbe21efd7c6a14"} Dec 05 10:55:13.076718 master-0 kubenswrapper[24928]: I1205 10:55:13.076678 24928 generic.go:334] "Generic (PLEG): container finished" podID="2934463a-d494-4735-a5d8-ad2c16aea065" containerID="2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9" exitCode=0 Dec 05 10:55:13.076718 master-0 kubenswrapper[24928]: I1205 10:55:13.076714 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4grl" event={"ID":"2934463a-d494-4735-a5d8-ad2c16aea065","Type":"ContainerDied","Data":"2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9"} Dec 05 10:55:13.076846 master-0 kubenswrapper[24928]: I1205 10:55:13.076736 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l4grl" event={"ID":"2934463a-d494-4735-a5d8-ad2c16aea065","Type":"ContainerDied","Data":"1f39e81a5d922f5c8f7636d4fc0f9b6b55b79a2ac69769e2833a8169baa5278c"} Dec 05 10:55:13.076846 master-0 kubenswrapper[24928]: I1205 10:55:13.076752 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l4grl" Dec 05 10:55:13.076985 master-0 kubenswrapper[24928]: I1205 10:55:13.076756 24928 scope.go:117] "RemoveContainer" containerID="2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9" Dec 05 10:55:13.165408 master-0 kubenswrapper[24928]: I1205 10:55:13.165373 24928 scope.go:117] "RemoveContainer" containerID="555f9686aee99474ecd7968cbfef0667db2d656f174e224a0a8651e26cd15b97" Dec 05 10:55:13.180263 master-0 kubenswrapper[24928]: I1205 10:55:13.180197 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4grl"] Dec 05 10:55:13.185814 master-0 kubenswrapper[24928]: I1205 10:55:13.185734 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l4grl"] Dec 05 10:55:13.202983 master-0 kubenswrapper[24928]: I1205 10:55:13.202739 24928 scope.go:117] "RemoveContainer" containerID="b4840415e898c8b5a1f111906d738ae0935737c840e6761d37b4ebefe81f411a" Dec 05 10:55:13.219723 master-0 kubenswrapper[24928]: I1205 10:55:13.219638 24928 scope.go:117] "RemoveContainer" containerID="2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9" Dec 05 10:55:13.222997 master-0 kubenswrapper[24928]: E1205 10:55:13.220988 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9\": container with ID starting with 2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9 not found: ID does not exist" containerID="2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9" Dec 05 10:55:13.222997 master-0 kubenswrapper[24928]: I1205 10:55:13.221051 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9"} err="failed to get container status \"2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9\": rpc error: code = NotFound desc = could not find container \"2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9\": container with ID starting with 2f30fea8b885e18dc6705faa2ea688fb1ad4e871c7b6efd0eec03c4659eb67d9 not found: ID does not exist" Dec 05 10:55:13.222997 master-0 kubenswrapper[24928]: I1205 10:55:13.221086 24928 scope.go:117] "RemoveContainer" containerID="555f9686aee99474ecd7968cbfef0667db2d656f174e224a0a8651e26cd15b97" Dec 05 10:55:13.223767 master-0 kubenswrapper[24928]: E1205 10:55:13.223086 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"555f9686aee99474ecd7968cbfef0667db2d656f174e224a0a8651e26cd15b97\": container with ID starting with 555f9686aee99474ecd7968cbfef0667db2d656f174e224a0a8651e26cd15b97 not found: ID does not exist" containerID="555f9686aee99474ecd7968cbfef0667db2d656f174e224a0a8651e26cd15b97" Dec 05 10:55:13.223767 master-0 kubenswrapper[24928]: I1205 10:55:13.223111 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"555f9686aee99474ecd7968cbfef0667db2d656f174e224a0a8651e26cd15b97"} err="failed to get container status \"555f9686aee99474ecd7968cbfef0667db2d656f174e224a0a8651e26cd15b97\": rpc error: code = NotFound desc = could not find container \"555f9686aee99474ecd7968cbfef0667db2d656f174e224a0a8651e26cd15b97\": container with ID starting with 555f9686aee99474ecd7968cbfef0667db2d656f174e224a0a8651e26cd15b97 not found: ID does not exist" Dec 05 10:55:13.223767 master-0 kubenswrapper[24928]: I1205 10:55:13.223131 24928 scope.go:117] "RemoveContainer" containerID="b4840415e898c8b5a1f111906d738ae0935737c840e6761d37b4ebefe81f411a" Dec 05 10:55:13.223767 master-0 kubenswrapper[24928]: E1205 10:55:13.223493 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4840415e898c8b5a1f111906d738ae0935737c840e6761d37b4ebefe81f411a\": container with ID starting with b4840415e898c8b5a1f111906d738ae0935737c840e6761d37b4ebefe81f411a not found: ID does not exist" containerID="b4840415e898c8b5a1f111906d738ae0935737c840e6761d37b4ebefe81f411a" Dec 05 10:55:13.223767 master-0 kubenswrapper[24928]: I1205 10:55:13.223547 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4840415e898c8b5a1f111906d738ae0935737c840e6761d37b4ebefe81f411a"} err="failed to get container status \"b4840415e898c8b5a1f111906d738ae0935737c840e6761d37b4ebefe81f411a\": rpc error: code = NotFound desc = could not find container \"b4840415e898c8b5a1f111906d738ae0935737c840e6761d37b4ebefe81f411a\": container with ID starting with b4840415e898c8b5a1f111906d738ae0935737c840e6761d37b4ebefe81f411a not found: ID does not exist" Dec 05 10:55:13.620472 master-0 kubenswrapper[24928]: I1205 10:55:13.605912 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:55:13.696801 master-0 kubenswrapper[24928]: I1205 10:55:13.696728 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c2j5\" (UniqueName: \"kubernetes.io/projected/8a54075d-4cba-439e-96c0-db16e991d658-kube-api-access-5c2j5\") pod \"8a54075d-4cba-439e-96c0-db16e991d658\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " Dec 05 10:55:13.697050 master-0 kubenswrapper[24928]: I1205 10:55:13.696865 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-utilities\") pod \"8a54075d-4cba-439e-96c0-db16e991d658\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " Dec 05 10:55:13.697469 master-0 kubenswrapper[24928]: I1205 10:55:13.697405 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-catalog-content\") pod \"8a54075d-4cba-439e-96c0-db16e991d658\" (UID: \"8a54075d-4cba-439e-96c0-db16e991d658\") " Dec 05 10:55:13.697982 master-0 kubenswrapper[24928]: I1205 10:55:13.697937 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-utilities" (OuterVolumeSpecName: "utilities") pod "8a54075d-4cba-439e-96c0-db16e991d658" (UID: "8a54075d-4cba-439e-96c0-db16e991d658"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:55:13.698095 master-0 kubenswrapper[24928]: I1205 10:55:13.698063 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:13.699698 master-0 kubenswrapper[24928]: I1205 10:55:13.699638 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a54075d-4cba-439e-96c0-db16e991d658-kube-api-access-5c2j5" (OuterVolumeSpecName: "kube-api-access-5c2j5") pod "8a54075d-4cba-439e-96c0-db16e991d658" (UID: "8a54075d-4cba-439e-96c0-db16e991d658"). InnerVolumeSpecName "kube-api-access-5c2j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:55:13.732175 master-0 kubenswrapper[24928]: I1205 10:55:13.732095 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74977ddd8b-dkrkh_629d199b-36e5-4c96-b58f-24314c4f2f58/console/0.log" Dec 05 10:55:13.732357 master-0 kubenswrapper[24928]: I1205 10:55:13.732204 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:55:13.800800 master-0 kubenswrapper[24928]: I1205 10:55:13.800733 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c2j5\" (UniqueName: \"kubernetes.io/projected/8a54075d-4cba-439e-96c0-db16e991d658-kube-api-access-5c2j5\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:13.815381 master-0 kubenswrapper[24928]: I1205 10:55:13.812335 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a54075d-4cba-439e-96c0-db16e991d658" (UID: "8a54075d-4cba-439e-96c0-db16e991d658"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:55:13.902181 master-0 kubenswrapper[24928]: I1205 10:55:13.902061 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-oauth-config\") pod \"629d199b-36e5-4c96-b58f-24314c4f2f58\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " Dec 05 10:55:13.902484 master-0 kubenswrapper[24928]: I1205 10:55:13.902235 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-trusted-ca-bundle\") pod \"629d199b-36e5-4c96-b58f-24314c4f2f58\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " Dec 05 10:55:13.902484 master-0 kubenswrapper[24928]: I1205 10:55:13.902285 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-oauth-serving-cert\") pod \"629d199b-36e5-4c96-b58f-24314c4f2f58\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " Dec 05 10:55:13.902484 master-0 kubenswrapper[24928]: I1205 10:55:13.902316 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-service-ca\") pod \"629d199b-36e5-4c96-b58f-24314c4f2f58\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " Dec 05 10:55:13.902484 master-0 kubenswrapper[24928]: I1205 10:55:13.902354 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw5lx\" (UniqueName: \"kubernetes.io/projected/629d199b-36e5-4c96-b58f-24314c4f2f58-kube-api-access-tw5lx\") pod \"629d199b-36e5-4c96-b58f-24314c4f2f58\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " Dec 05 10:55:13.902484 master-0 kubenswrapper[24928]: I1205 10:55:13.902414 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-serving-cert\") pod \"629d199b-36e5-4c96-b58f-24314c4f2f58\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " Dec 05 10:55:13.902732 master-0 kubenswrapper[24928]: I1205 10:55:13.902500 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-console-config\") pod \"629d199b-36e5-4c96-b58f-24314c4f2f58\" (UID: \"629d199b-36e5-4c96-b58f-24314c4f2f58\") " Dec 05 10:55:13.902845 master-0 kubenswrapper[24928]: I1205 10:55:13.902801 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a54075d-4cba-439e-96c0-db16e991d658-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:13.903009 master-0 kubenswrapper[24928]: I1205 10:55:13.902959 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "629d199b-36e5-4c96-b58f-24314c4f2f58" (UID: "629d199b-36e5-4c96-b58f-24314c4f2f58"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:55:13.903057 master-0 kubenswrapper[24928]: I1205 10:55:13.902990 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-service-ca" (OuterVolumeSpecName: "service-ca") pod "629d199b-36e5-4c96-b58f-24314c4f2f58" (UID: "629d199b-36e5-4c96-b58f-24314c4f2f58"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:55:13.903057 master-0 kubenswrapper[24928]: I1205 10:55:13.902945 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "629d199b-36e5-4c96-b58f-24314c4f2f58" (UID: "629d199b-36e5-4c96-b58f-24314c4f2f58"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:55:13.903316 master-0 kubenswrapper[24928]: I1205 10:55:13.903270 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-console-config" (OuterVolumeSpecName: "console-config") pod "629d199b-36e5-4c96-b58f-24314c4f2f58" (UID: "629d199b-36e5-4c96-b58f-24314c4f2f58"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:55:13.905554 master-0 kubenswrapper[24928]: I1205 10:55:13.905506 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "629d199b-36e5-4c96-b58f-24314c4f2f58" (UID: "629d199b-36e5-4c96-b58f-24314c4f2f58"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:55:13.905746 master-0 kubenswrapper[24928]: I1205 10:55:13.905722 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "629d199b-36e5-4c96-b58f-24314c4f2f58" (UID: "629d199b-36e5-4c96-b58f-24314c4f2f58"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:55:13.906572 master-0 kubenswrapper[24928]: I1205 10:55:13.906460 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/629d199b-36e5-4c96-b58f-24314c4f2f58-kube-api-access-tw5lx" (OuterVolumeSpecName: "kube-api-access-tw5lx") pod "629d199b-36e5-4c96-b58f-24314c4f2f58" (UID: "629d199b-36e5-4c96-b58f-24314c4f2f58"). InnerVolumeSpecName "kube-api-access-tw5lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:55:14.006407 master-0 kubenswrapper[24928]: I1205 10:55:14.005284 24928 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:14.006407 master-0 kubenswrapper[24928]: I1205 10:55:14.006201 24928 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-console-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:14.006407 master-0 kubenswrapper[24928]: I1205 10:55:14.006218 24928 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/629d199b-36e5-4c96-b58f-24314c4f2f58-console-oauth-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:14.006407 master-0 kubenswrapper[24928]: I1205 10:55:14.006231 24928 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-trusted-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:14.006407 master-0 kubenswrapper[24928]: I1205 10:55:14.006244 24928 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-oauth-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:14.006407 master-0 kubenswrapper[24928]: I1205 10:55:14.006260 24928 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/629d199b-36e5-4c96-b58f-24314c4f2f58-service-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:14.006407 master-0 kubenswrapper[24928]: I1205 10:55:14.006273 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw5lx\" (UniqueName: \"kubernetes.io/projected/629d199b-36e5-4c96-b58f-24314c4f2f58-kube-api-access-tw5lx\") on node \"master-0\" DevicePath \"\"" Dec 05 10:55:14.085531 master-0 kubenswrapper[24928]: I1205 10:55:14.085472 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-74977ddd8b-dkrkh_629d199b-36e5-4c96-b58f-24314c4f2f58/console/0.log" Dec 05 10:55:14.085739 master-0 kubenswrapper[24928]: I1205 10:55:14.085581 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-74977ddd8b-dkrkh" event={"ID":"629d199b-36e5-4c96-b58f-24314c4f2f58","Type":"ContainerDied","Data":"ba2dee4c8f6ea28a581d1b8366715076c17c0d605303d3209026cf7512384aca"} Dec 05 10:55:14.085739 master-0 kubenswrapper[24928]: I1205 10:55:14.085633 24928 scope.go:117] "RemoveContainer" containerID="f1a948781cc195931bc78694a8e65f57a216949ec87c08ca993d58ff6b6fbbd1" Dec 05 10:55:14.085739 master-0 kubenswrapper[24928]: I1205 10:55:14.085643 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-74977ddd8b-dkrkh" Dec 05 10:55:14.088314 master-0 kubenswrapper[24928]: I1205 10:55:14.088266 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8pb58" event={"ID":"8a54075d-4cba-439e-96c0-db16e991d658","Type":"ContainerDied","Data":"bace8793d0c45eb8ada78dde17ec3e9d888c36fef2e3c8dc7397b9a47a136c43"} Dec 05 10:55:14.088314 master-0 kubenswrapper[24928]: I1205 10:55:14.088279 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8pb58" Dec 05 10:55:14.101084 master-0 kubenswrapper[24928]: I1205 10:55:14.101035 24928 scope.go:117] "RemoveContainer" containerID="af68b8278694217e9435580393ef2485614c877c89eaa1ccd5cbe21efd7c6a14" Dec 05 10:55:14.120248 master-0 kubenswrapper[24928]: I1205 10:55:14.119946 24928 scope.go:117] "RemoveContainer" containerID="5328f8133a36f39e3b380fc3497d03e68bbc44cc892da222cfc20c5c6eda1fe7" Dec 05 10:55:14.137022 master-0 kubenswrapper[24928]: I1205 10:55:14.134740 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8pb58"] Dec 05 10:55:14.143235 master-0 kubenswrapper[24928]: I1205 10:55:14.143162 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8pb58"] Dec 05 10:55:14.154016 master-0 kubenswrapper[24928]: I1205 10:55:14.153933 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-74977ddd8b-dkrkh"] Dec 05 10:55:14.157248 master-0 kubenswrapper[24928]: I1205 10:55:14.157185 24928 scope.go:117] "RemoveContainer" containerID="0ee7dee1876c3ed1dcc8f819d2ac65a461808f03de47099ca1a137bd407fb7b7" Dec 05 10:55:14.158903 master-0 kubenswrapper[24928]: I1205 10:55:14.158835 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-74977ddd8b-dkrkh"] Dec 05 10:55:14.216738 master-0 kubenswrapper[24928]: I1205 10:55:14.216673 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2934463a-d494-4735-a5d8-ad2c16aea065" path="/var/lib/kubelet/pods/2934463a-d494-4735-a5d8-ad2c16aea065/volumes" Dec 05 10:55:14.217581 master-0 kubenswrapper[24928]: I1205 10:55:14.217556 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="629d199b-36e5-4c96-b58f-24314c4f2f58" path="/var/lib/kubelet/pods/629d199b-36e5-4c96-b58f-24314c4f2f58/volumes" Dec 05 10:55:14.219010 master-0 kubenswrapper[24928]: I1205 10:55:14.218165 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a54075d-4cba-439e-96c0-db16e991d658" path="/var/lib/kubelet/pods/8a54075d-4cba-439e-96c0-db16e991d658/volumes" Dec 05 10:55:15.396011 master-0 kubenswrapper[24928]: I1205 10:55:15.395896 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:55:15.403366 master-0 kubenswrapper[24928]: I1205 10:55:15.403281 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-64494f74c5-sqgmf" Dec 05 10:55:35.188486 master-0 kubenswrapper[24928]: I1205 10:55:35.188404 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:55:35.225245 master-0 kubenswrapper[24928]: I1205 10:55:35.225181 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:55:35.276260 master-0 kubenswrapper[24928]: I1205 10:55:35.276200 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Dec 05 10:56:05.484683 master-0 kubenswrapper[24928]: I1205 10:56:05.484620 24928 generic.go:334] "Generic (PLEG): container finished" podID="3aa9a063-322b-4dc6-a724-05a66026160b" containerID="20fd7ecb75706506f53b22d6ef8c2f7cefab03481f4f534c87bae4cf23aa6438" exitCode=0 Dec 05 10:56:05.485150 master-0 kubenswrapper[24928]: I1205 10:56:05.484693 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" event={"ID":"3aa9a063-322b-4dc6-a724-05a66026160b","Type":"ContainerDied","Data":"20fd7ecb75706506f53b22d6ef8c2f7cefab03481f4f534c87bae4cf23aa6438"} Dec 05 10:56:05.750620 master-0 kubenswrapper[24928]: I1205 10:56:05.750563 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:56:05.842454 master-0 kubenswrapper[24928]: I1205 10:56:05.842338 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle\") pod \"3aa9a063-322b-4dc6-a724-05a66026160b\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " Dec 05 10:56:05.842454 master-0 kubenswrapper[24928]: I1205 10:56:05.842467 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rlts\" (UniqueName: \"kubernetes.io/projected/3aa9a063-322b-4dc6-a724-05a66026160b-kube-api-access-5rlts\") pod \"3aa9a063-322b-4dc6-a724-05a66026160b\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " Dec 05 10:56:05.842763 master-0 kubenswrapper[24928]: I1205 10:56:05.842529 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle\") pod \"3aa9a063-322b-4dc6-a724-05a66026160b\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " Dec 05 10:56:05.842763 master-0 kubenswrapper[24928]: I1205 10:56:05.842579 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/3aa9a063-322b-4dc6-a724-05a66026160b-audit-log\") pod \"3aa9a063-322b-4dc6-a724-05a66026160b\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " Dec 05 10:56:05.842763 master-0 kubenswrapper[24928]: I1205 10:56:05.842662 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles\") pod \"3aa9a063-322b-4dc6-a724-05a66026160b\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " Dec 05 10:56:05.842897 master-0 kubenswrapper[24928]: I1205 10:56:05.842771 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs\") pod \"3aa9a063-322b-4dc6-a724-05a66026160b\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " Dec 05 10:56:05.842897 master-0 kubenswrapper[24928]: I1205 10:56:05.842831 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls\") pod \"3aa9a063-322b-4dc6-a724-05a66026160b\" (UID: \"3aa9a063-322b-4dc6-a724-05a66026160b\") " Dec 05 10:56:05.843151 master-0 kubenswrapper[24928]: I1205 10:56:05.843096 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3aa9a063-322b-4dc6-a724-05a66026160b-audit-log" (OuterVolumeSpecName: "audit-log") pod "3aa9a063-322b-4dc6-a724-05a66026160b" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b"). InnerVolumeSpecName "audit-log". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:56:05.843286 master-0 kubenswrapper[24928]: I1205 10:56:05.843231 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles" (OuterVolumeSpecName: "metrics-server-audit-profiles") pod "3aa9a063-322b-4dc6-a724-05a66026160b" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b"). InnerVolumeSpecName "metrics-server-audit-profiles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:56:05.843450 master-0 kubenswrapper[24928]: I1205 10:56:05.843397 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle" (OuterVolumeSpecName: "configmap-kubelet-serving-ca-bundle") pod "3aa9a063-322b-4dc6-a724-05a66026160b" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b"). InnerVolumeSpecName "configmap-kubelet-serving-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:56:05.844045 master-0 kubenswrapper[24928]: I1205 10:56:05.844008 24928 reconciler_common.go:293] "Volume detached for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/3aa9a063-322b-4dc6-a724-05a66026160b-audit-log\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:05.844123 master-0 kubenswrapper[24928]: I1205 10:56:05.844043 24928 reconciler_common.go:293] "Volume detached for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-metrics-server-audit-profiles\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:05.844123 master-0 kubenswrapper[24928]: I1205 10:56:05.844062 24928 reconciler_common.go:293] "Volume detached for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3aa9a063-322b-4dc6-a724-05a66026160b-configmap-kubelet-serving-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:05.847673 master-0 kubenswrapper[24928]: I1205 10:56:05.846064 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls" (OuterVolumeSpecName: "secret-metrics-server-tls") pod "3aa9a063-322b-4dc6-a724-05a66026160b" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b"). InnerVolumeSpecName "secret-metrics-server-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:56:05.847673 master-0 kubenswrapper[24928]: I1205 10:56:05.846119 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs" (OuterVolumeSpecName: "secret-metrics-client-certs") pod "3aa9a063-322b-4dc6-a724-05a66026160b" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b"). InnerVolumeSpecName "secret-metrics-client-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:56:05.847673 master-0 kubenswrapper[24928]: I1205 10:56:05.846776 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle" (OuterVolumeSpecName: "client-ca-bundle") pod "3aa9a063-322b-4dc6-a724-05a66026160b" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b"). InnerVolumeSpecName "client-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:56:05.849669 master-0 kubenswrapper[24928]: I1205 10:56:05.849634 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa9a063-322b-4dc6-a724-05a66026160b-kube-api-access-5rlts" (OuterVolumeSpecName: "kube-api-access-5rlts") pod "3aa9a063-322b-4dc6-a724-05a66026160b" (UID: "3aa9a063-322b-4dc6-a724-05a66026160b"). InnerVolumeSpecName "kube-api-access-5rlts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:56:05.945364 master-0 kubenswrapper[24928]: I1205 10:56:05.945246 24928 reconciler_common.go:293] "Volume detached for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-client-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:05.945364 master-0 kubenswrapper[24928]: I1205 10:56:05.945297 24928 reconciler_common.go:293] "Volume detached for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-secret-metrics-server-tls\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:05.945364 master-0 kubenswrapper[24928]: I1205 10:56:05.945312 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rlts\" (UniqueName: \"kubernetes.io/projected/3aa9a063-322b-4dc6-a724-05a66026160b-kube-api-access-5rlts\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:05.945364 master-0 kubenswrapper[24928]: I1205 10:56:05.945330 24928 reconciler_common.go:293] "Volume detached for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aa9a063-322b-4dc6-a724-05a66026160b-client-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:05.980131 master-0 kubenswrapper[24928]: I1205 10:56:05.979969 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-75dfb65779-zgfwv"] Dec 05 10:56:05.980335 master-0 kubenswrapper[24928]: E1205 10:56:05.980295 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" containerName="extract-content" Dec 05 10:56:05.980335 master-0 kubenswrapper[24928]: I1205 10:56:05.980308 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" containerName="extract-content" Dec 05 10:56:05.980335 master-0 kubenswrapper[24928]: E1205 10:56:05.980319 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629d199b-36e5-4c96-b58f-24314c4f2f58" containerName="console" Dec 05 10:56:05.980335 master-0 kubenswrapper[24928]: I1205 10:56:05.980325 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="629d199b-36e5-4c96-b58f-24314c4f2f58" containerName="console" Dec 05 10:56:05.980335 master-0 kubenswrapper[24928]: E1205 10:56:05.980336 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2934463a-d494-4735-a5d8-ad2c16aea065" containerName="extract-content" Dec 05 10:56:05.980335 master-0 kubenswrapper[24928]: I1205 10:56:05.980342 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="2934463a-d494-4735-a5d8-ad2c16aea065" containerName="extract-content" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: E1205 10:56:05.980353 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25d480c2-7108-450d-adc7-c56e0b5be0c8" containerName="extract-utilities" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: I1205 10:56:05.980359 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="25d480c2-7108-450d-adc7-c56e0b5be0c8" containerName="extract-utilities" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: E1205 10:56:05.980377 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2934463a-d494-4735-a5d8-ad2c16aea065" containerName="extract-utilities" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: I1205 10:56:05.980385 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="2934463a-d494-4735-a5d8-ad2c16aea065" containerName="extract-utilities" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: E1205 10:56:05.980398 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a54075d-4cba-439e-96c0-db16e991d658" containerName="registry-server" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: I1205 10:56:05.980405 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a54075d-4cba-439e-96c0-db16e991d658" containerName="registry-server" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: E1205 10:56:05.980446 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" containerName="registry-server" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: I1205 10:56:05.980453 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" containerName="registry-server" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: E1205 10:56:05.980464 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a54075d-4cba-439e-96c0-db16e991d658" containerName="extract-utilities" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: I1205 10:56:05.980470 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a54075d-4cba-439e-96c0-db16e991d658" containerName="extract-utilities" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: E1205 10:56:05.980481 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa9a063-322b-4dc6-a724-05a66026160b" containerName="metrics-server" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: I1205 10:56:05.980486 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa9a063-322b-4dc6-a724-05a66026160b" containerName="metrics-server" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: E1205 10:56:05.980498 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a54075d-4cba-439e-96c0-db16e991d658" containerName="extract-content" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: I1205 10:56:05.980504 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a54075d-4cba-439e-96c0-db16e991d658" containerName="extract-content" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: E1205 10:56:05.980513 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25d480c2-7108-450d-adc7-c56e0b5be0c8" containerName="registry-server" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: I1205 10:56:05.980519 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="25d480c2-7108-450d-adc7-c56e0b5be0c8" containerName="registry-server" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: E1205 10:56:05.980536 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" containerName="extract-utilities" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: I1205 10:56:05.980542 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" containerName="extract-utilities" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: E1205 10:56:05.980554 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2934463a-d494-4735-a5d8-ad2c16aea065" containerName="registry-server" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: I1205 10:56:05.980559 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="2934463a-d494-4735-a5d8-ad2c16aea065" containerName="registry-server" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: E1205 10:56:05.980568 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25d480c2-7108-450d-adc7-c56e0b5be0c8" containerName="extract-content" Dec 05 10:56:05.980613 master-0 kubenswrapper[24928]: I1205 10:56:05.980574 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="25d480c2-7108-450d-adc7-c56e0b5be0c8" containerName="extract-content" Dec 05 10:56:05.981316 master-0 kubenswrapper[24928]: I1205 10:56:05.980683 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a54075d-4cba-439e-96c0-db16e991d658" containerName="registry-server" Dec 05 10:56:05.981316 master-0 kubenswrapper[24928]: I1205 10:56:05.980699 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa9a063-322b-4dc6-a724-05a66026160b" containerName="metrics-server" Dec 05 10:56:05.981316 master-0 kubenswrapper[24928]: I1205 10:56:05.980714 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="629d199b-36e5-4c96-b58f-24314c4f2f58" containerName="console" Dec 05 10:56:05.981316 master-0 kubenswrapper[24928]: I1205 10:56:05.980728 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="241003b6-4aa6-4ed5-8bdd-46f2d30662fd" containerName="registry-server" Dec 05 10:56:05.981316 master-0 kubenswrapper[24928]: I1205 10:56:05.980737 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="25d480c2-7108-450d-adc7-c56e0b5be0c8" containerName="registry-server" Dec 05 10:56:05.981316 master-0 kubenswrapper[24928]: I1205 10:56:05.980760 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="2934463a-d494-4735-a5d8-ad2c16aea065" containerName="registry-server" Dec 05 10:56:05.981316 master-0 kubenswrapper[24928]: I1205 10:56:05.981213 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.016240 master-0 kubenswrapper[24928]: I1205 10:56:06.016156 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75dfb65779-zgfwv"] Dec 05 10:56:06.046546 master-0 kubenswrapper[24928]: I1205 10:56:06.046409 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-service-ca\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.046888 master-0 kubenswrapper[24928]: I1205 10:56:06.046773 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-serving-cert\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.046888 master-0 kubenswrapper[24928]: I1205 10:56:06.046867 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-oauth-serving-cert\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.047097 master-0 kubenswrapper[24928]: I1205 10:56:06.047061 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-986m9\" (UniqueName: \"kubernetes.io/projected/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-kube-api-access-986m9\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.047156 master-0 kubenswrapper[24928]: I1205 10:56:06.047134 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-trusted-ca-bundle\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.047394 master-0 kubenswrapper[24928]: I1205 10:56:06.047358 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-oauth-config\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.047674 master-0 kubenswrapper[24928]: I1205 10:56:06.047528 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-config\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.150569 master-0 kubenswrapper[24928]: I1205 10:56:06.150346 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-oauth-serving-cert\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.150569 master-0 kubenswrapper[24928]: I1205 10:56:06.150444 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-986m9\" (UniqueName: \"kubernetes.io/projected/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-kube-api-access-986m9\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.150997 master-0 kubenswrapper[24928]: I1205 10:56:06.150678 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-trusted-ca-bundle\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.151058 master-0 kubenswrapper[24928]: I1205 10:56:06.151027 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-oauth-config\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.152234 master-0 kubenswrapper[24928]: I1205 10:56:06.151130 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-config\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.152325 master-0 kubenswrapper[24928]: I1205 10:56:06.151713 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-oauth-serving-cert\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.152325 master-0 kubenswrapper[24928]: I1205 10:56:06.152155 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-config\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.152473 master-0 kubenswrapper[24928]: I1205 10:56:06.152415 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-service-ca\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.154499 master-0 kubenswrapper[24928]: I1205 10:56:06.152558 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-trusted-ca-bundle\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.154499 master-0 kubenswrapper[24928]: I1205 10:56:06.152661 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-serving-cert\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.154499 master-0 kubenswrapper[24928]: I1205 10:56:06.153298 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-service-ca\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.157954 master-0 kubenswrapper[24928]: I1205 10:56:06.157833 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-oauth-config\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.160189 master-0 kubenswrapper[24928]: I1205 10:56:06.158817 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-serving-cert\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.169347 master-0 kubenswrapper[24928]: I1205 10:56:06.169289 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-986m9\" (UniqueName: \"kubernetes.io/projected/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-kube-api-access-986m9\") pod \"console-75dfb65779-zgfwv\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.302657 master-0 kubenswrapper[24928]: I1205 10:56:06.302576 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:06.537038 master-0 kubenswrapper[24928]: I1205 10:56:06.536962 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" event={"ID":"3aa9a063-322b-4dc6-a724-05a66026160b","Type":"ContainerDied","Data":"9dab351b1866858c81dfc4a1e5ed04a7dcaa8ee50017ec3f0b131de9c613ce86"} Dec 05 10:56:06.537038 master-0 kubenswrapper[24928]: I1205 10:56:06.537038 24928 scope.go:117] "RemoveContainer" containerID="20fd7ecb75706506f53b22d6ef8c2f7cefab03481f4f534c87bae4cf23aa6438" Dec 05 10:56:06.537684 master-0 kubenswrapper[24928]: I1205 10:56:06.537156 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-7c46d76dff-z8d8z" Dec 05 10:56:06.562489 master-0 kubenswrapper[24928]: I1205 10:56:06.561388 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-monitoring/metrics-server-7c46d76dff-z8d8z"] Dec 05 10:56:06.565921 master-0 kubenswrapper[24928]: I1205 10:56:06.565859 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-monitoring/metrics-server-7c46d76dff-z8d8z"] Dec 05 10:56:06.861332 master-0 kubenswrapper[24928]: W1205 10:56:06.861123 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda92cfe85_6e5d_4307_bb0d_4bc7dfa2be6e.slice/crio-0bc51a3efdd9870480dec9ce3757dff7dac46efb97f0b577b15dbdbb26d34a55 WatchSource:0}: Error finding container 0bc51a3efdd9870480dec9ce3757dff7dac46efb97f0b577b15dbdbb26d34a55: Status 404 returned error can't find the container with id 0bc51a3efdd9870480dec9ce3757dff7dac46efb97f0b577b15dbdbb26d34a55 Dec 05 10:56:06.868158 master-0 kubenswrapper[24928]: I1205 10:56:06.868082 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75dfb65779-zgfwv"] Dec 05 10:56:07.546964 master-0 kubenswrapper[24928]: I1205 10:56:07.546896 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75dfb65779-zgfwv" event={"ID":"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e","Type":"ContainerStarted","Data":"d9af2b69c83d05345ecddb38e578653ae88426a35a9767f27c9efb807c0762b1"} Dec 05 10:56:07.547629 master-0 kubenswrapper[24928]: I1205 10:56:07.546963 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75dfb65779-zgfwv" event={"ID":"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e","Type":"ContainerStarted","Data":"0bc51a3efdd9870480dec9ce3757dff7dac46efb97f0b577b15dbdbb26d34a55"} Dec 05 10:56:07.592403 master-0 kubenswrapper[24928]: I1205 10:56:07.591884 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-75dfb65779-zgfwv" podStartSLOduration=2.591857228 podStartE2EDuration="2.591857228s" podCreationTimestamp="2025-12-05 10:56:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:56:07.587993368 +0000 UTC m=+527.591187219" watchObservedRunningTime="2025-12-05 10:56:07.591857228 +0000 UTC m=+527.595051089" Dec 05 10:56:08.212668 master-0 kubenswrapper[24928]: I1205 10:56:08.212595 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aa9a063-322b-4dc6-a724-05a66026160b" path="/var/lib/kubelet/pods/3aa9a063-322b-4dc6-a724-05a66026160b/volumes" Dec 05 10:56:16.303273 master-0 kubenswrapper[24928]: I1205 10:56:16.303168 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:16.303273 master-0 kubenswrapper[24928]: I1205 10:56:16.303245 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:16.308029 master-0 kubenswrapper[24928]: I1205 10:56:16.307921 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:16.615337 master-0 kubenswrapper[24928]: I1205 10:56:16.615218 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:56:16.688790 master-0 kubenswrapper[24928]: I1205 10:56:16.688712 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-86b5fdbff8-6l4nn"] Dec 05 10:56:41.743664 master-0 kubenswrapper[24928]: I1205 10:56:41.743555 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-86b5fdbff8-6l4nn" podUID="be90caa2-11f2-41f3-af4b-2df058459e12" containerName="console" containerID="cri-o://861de06c2bcb98334d269cb4c41d1465a0dcd05c72e52912435a598cd19bf542" gracePeriod=15 Dec 05 10:56:42.825988 master-0 kubenswrapper[24928]: I1205 10:56:42.825935 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-86b5fdbff8-6l4nn_be90caa2-11f2-41f3-af4b-2df058459e12/console/0.log" Dec 05 10:56:42.825988 master-0 kubenswrapper[24928]: I1205 10:56:42.825984 24928 generic.go:334] "Generic (PLEG): container finished" podID="be90caa2-11f2-41f3-af4b-2df058459e12" containerID="861de06c2bcb98334d269cb4c41d1465a0dcd05c72e52912435a598cd19bf542" exitCode=2 Dec 05 10:56:42.826557 master-0 kubenswrapper[24928]: I1205 10:56:42.826018 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86b5fdbff8-6l4nn" event={"ID":"be90caa2-11f2-41f3-af4b-2df058459e12","Type":"ContainerDied","Data":"861de06c2bcb98334d269cb4c41d1465a0dcd05c72e52912435a598cd19bf542"} Dec 05 10:56:42.983504 master-0 kubenswrapper[24928]: I1205 10:56:42.983400 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-86b5fdbff8-6l4nn_be90caa2-11f2-41f3-af4b-2df058459e12/console/0.log" Dec 05 10:56:42.983765 master-0 kubenswrapper[24928]: I1205 10:56:42.983536 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:56:43.160930 master-0 kubenswrapper[24928]: I1205 10:56:43.160854 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-service-ca\") pod \"be90caa2-11f2-41f3-af4b-2df058459e12\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " Dec 05 10:56:43.161409 master-0 kubenswrapper[24928]: I1205 10:56:43.160975 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-serving-cert\") pod \"be90caa2-11f2-41f3-af4b-2df058459e12\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " Dec 05 10:56:43.161409 master-0 kubenswrapper[24928]: I1205 10:56:43.161079 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-oauth-serving-cert\") pod \"be90caa2-11f2-41f3-af4b-2df058459e12\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " Dec 05 10:56:43.161409 master-0 kubenswrapper[24928]: I1205 10:56:43.161136 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zhptk\" (UniqueName: \"kubernetes.io/projected/be90caa2-11f2-41f3-af4b-2df058459e12-kube-api-access-zhptk\") pod \"be90caa2-11f2-41f3-af4b-2df058459e12\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " Dec 05 10:56:43.161409 master-0 kubenswrapper[24928]: I1205 10:56:43.161220 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-oauth-config\") pod \"be90caa2-11f2-41f3-af4b-2df058459e12\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " Dec 05 10:56:43.161409 master-0 kubenswrapper[24928]: I1205 10:56:43.161256 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-trusted-ca-bundle\") pod \"be90caa2-11f2-41f3-af4b-2df058459e12\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " Dec 05 10:56:43.161409 master-0 kubenswrapper[24928]: I1205 10:56:43.161276 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-console-config\") pod \"be90caa2-11f2-41f3-af4b-2df058459e12\" (UID: \"be90caa2-11f2-41f3-af4b-2df058459e12\") " Dec 05 10:56:43.162175 master-0 kubenswrapper[24928]: I1205 10:56:43.162116 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-console-config" (OuterVolumeSpecName: "console-config") pod "be90caa2-11f2-41f3-af4b-2df058459e12" (UID: "be90caa2-11f2-41f3-af4b-2df058459e12"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:56:43.163082 master-0 kubenswrapper[24928]: I1205 10:56:43.163052 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "be90caa2-11f2-41f3-af4b-2df058459e12" (UID: "be90caa2-11f2-41f3-af4b-2df058459e12"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:56:43.163221 master-0 kubenswrapper[24928]: I1205 10:56:43.163068 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-service-ca" (OuterVolumeSpecName: "service-ca") pod "be90caa2-11f2-41f3-af4b-2df058459e12" (UID: "be90caa2-11f2-41f3-af4b-2df058459e12"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:56:43.163376 master-0 kubenswrapper[24928]: I1205 10:56:43.163090 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "be90caa2-11f2-41f3-af4b-2df058459e12" (UID: "be90caa2-11f2-41f3-af4b-2df058459e12"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:56:43.165782 master-0 kubenswrapper[24928]: I1205 10:56:43.165718 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "be90caa2-11f2-41f3-af4b-2df058459e12" (UID: "be90caa2-11f2-41f3-af4b-2df058459e12"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:56:43.165782 master-0 kubenswrapper[24928]: I1205 10:56:43.165705 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "be90caa2-11f2-41f3-af4b-2df058459e12" (UID: "be90caa2-11f2-41f3-af4b-2df058459e12"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:56:43.165992 master-0 kubenswrapper[24928]: I1205 10:56:43.165892 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be90caa2-11f2-41f3-af4b-2df058459e12-kube-api-access-zhptk" (OuterVolumeSpecName: "kube-api-access-zhptk") pod "be90caa2-11f2-41f3-af4b-2df058459e12" (UID: "be90caa2-11f2-41f3-af4b-2df058459e12"). InnerVolumeSpecName "kube-api-access-zhptk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:56:43.264175 master-0 kubenswrapper[24928]: I1205 10:56:43.263573 24928 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-oauth-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:43.264175 master-0 kubenswrapper[24928]: I1205 10:56:43.263671 24928 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-trusted-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:43.264175 master-0 kubenswrapper[24928]: I1205 10:56:43.263682 24928 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-console-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:43.264175 master-0 kubenswrapper[24928]: I1205 10:56:43.263694 24928 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-service-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:43.264175 master-0 kubenswrapper[24928]: I1205 10:56:43.263703 24928 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/be90caa2-11f2-41f3-af4b-2df058459e12-console-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:43.264175 master-0 kubenswrapper[24928]: I1205 10:56:43.263712 24928 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/be90caa2-11f2-41f3-af4b-2df058459e12-oauth-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:43.264175 master-0 kubenswrapper[24928]: I1205 10:56:43.263721 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zhptk\" (UniqueName: \"kubernetes.io/projected/be90caa2-11f2-41f3-af4b-2df058459e12-kube-api-access-zhptk\") on node \"master-0\" DevicePath \"\"" Dec 05 10:56:43.836619 master-0 kubenswrapper[24928]: I1205 10:56:43.836530 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-86b5fdbff8-6l4nn_be90caa2-11f2-41f3-af4b-2df058459e12/console/0.log" Dec 05 10:56:43.837348 master-0 kubenswrapper[24928]: I1205 10:56:43.836639 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-86b5fdbff8-6l4nn" event={"ID":"be90caa2-11f2-41f3-af4b-2df058459e12","Type":"ContainerDied","Data":"db59570c8a8ddd33f58367b6f7074b7a3ec8bed27ec3277b0d41df81b8b48d2c"} Dec 05 10:56:43.837348 master-0 kubenswrapper[24928]: I1205 10:56:43.836708 24928 scope.go:117] "RemoveContainer" containerID="861de06c2bcb98334d269cb4c41d1465a0dcd05c72e52912435a598cd19bf542" Dec 05 10:56:43.837348 master-0 kubenswrapper[24928]: I1205 10:56:43.836715 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-86b5fdbff8-6l4nn" Dec 05 10:56:44.687824 master-0 kubenswrapper[24928]: I1205 10:56:44.687749 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-86b5fdbff8-6l4nn"] Dec 05 10:56:44.702773 master-0 kubenswrapper[24928]: I1205 10:56:44.702349 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-86b5fdbff8-6l4nn"] Dec 05 10:56:46.217612 master-0 kubenswrapper[24928]: I1205 10:56:46.217512 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be90caa2-11f2-41f3-af4b-2df058459e12" path="/var/lib/kubelet/pods/be90caa2-11f2-41f3-af4b-2df058459e12/volumes" Dec 05 10:56:57.797358 master-0 kubenswrapper[24928]: I1205 10:56:57.797246 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-d656f4996-kjkt5"] Dec 05 10:56:57.799450 master-0 kubenswrapper[24928]: E1205 10:56:57.797661 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be90caa2-11f2-41f3-af4b-2df058459e12" containerName="console" Dec 05 10:56:57.799450 master-0 kubenswrapper[24928]: I1205 10:56:57.797680 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="be90caa2-11f2-41f3-af4b-2df058459e12" containerName="console" Dec 05 10:56:57.799450 master-0 kubenswrapper[24928]: I1205 10:56:57.797862 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="be90caa2-11f2-41f3-af4b-2df058459e12" containerName="console" Dec 05 10:56:57.799450 master-0 kubenswrapper[24928]: I1205 10:56:57.798393 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.818272 master-0 kubenswrapper[24928]: I1205 10:56:57.817448 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-d656f4996-kjkt5"] Dec 05 10:56:57.892260 master-0 kubenswrapper[24928]: I1205 10:56:57.892204 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-service-ca\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.892510 master-0 kubenswrapper[24928]: I1205 10:56:57.892276 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-oauth-serving-cert\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.892510 master-0 kubenswrapper[24928]: I1205 10:56:57.892326 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-trusted-ca-bundle\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.892510 master-0 kubenswrapper[24928]: I1205 10:56:57.892354 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-serving-cert\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.892510 master-0 kubenswrapper[24928]: I1205 10:56:57.892453 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-oauth-config\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.892510 master-0 kubenswrapper[24928]: I1205 10:56:57.892488 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd7s5\" (UniqueName: \"kubernetes.io/projected/30aa818a-f5af-4468-9920-fb1f8d1cb811-kube-api-access-wd7s5\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.892739 master-0 kubenswrapper[24928]: I1205 10:56:57.892514 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-config\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.994213 master-0 kubenswrapper[24928]: I1205 10:56:57.994148 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-trusted-ca-bundle\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.994213 master-0 kubenswrapper[24928]: I1205 10:56:57.994209 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-serving-cert\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.994548 master-0 kubenswrapper[24928]: I1205 10:56:57.994282 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-oauth-config\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.994548 master-0 kubenswrapper[24928]: I1205 10:56:57.994308 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd7s5\" (UniqueName: \"kubernetes.io/projected/30aa818a-f5af-4468-9920-fb1f8d1cb811-kube-api-access-wd7s5\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.994548 master-0 kubenswrapper[24928]: I1205 10:56:57.994329 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-config\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.994548 master-0 kubenswrapper[24928]: I1205 10:56:57.994498 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-service-ca\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.994792 master-0 kubenswrapper[24928]: I1205 10:56:57.994613 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-oauth-serving-cert\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.995601 master-0 kubenswrapper[24928]: I1205 10:56:57.995264 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-config\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.995601 master-0 kubenswrapper[24928]: I1205 10:56:57.995586 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-oauth-serving-cert\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.996096 master-0 kubenswrapper[24928]: I1205 10:56:57.996069 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-trusted-ca-bundle\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.997199 master-0 kubenswrapper[24928]: I1205 10:56:57.996892 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-service-ca\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.999123 master-0 kubenswrapper[24928]: I1205 10:56:57.998747 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-serving-cert\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:57.999123 master-0 kubenswrapper[24928]: I1205 10:56:57.998954 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-oauth-config\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:58.012504 master-0 kubenswrapper[24928]: I1205 10:56:58.012455 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd7s5\" (UniqueName: \"kubernetes.io/projected/30aa818a-f5af-4468-9920-fb1f8d1cb811-kube-api-access-wd7s5\") pod \"console-d656f4996-kjkt5\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:58.115516 master-0 kubenswrapper[24928]: I1205 10:56:58.115353 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:56:58.543078 master-0 kubenswrapper[24928]: W1205 10:56:58.543018 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30aa818a_f5af_4468_9920_fb1f8d1cb811.slice/crio-545acad0541c1a9d6068ca98ebe068536bad5bac9c64a9ce82ec7c0d3d590c94 WatchSource:0}: Error finding container 545acad0541c1a9d6068ca98ebe068536bad5bac9c64a9ce82ec7c0d3d590c94: Status 404 returned error can't find the container with id 545acad0541c1a9d6068ca98ebe068536bad5bac9c64a9ce82ec7c0d3d590c94 Dec 05 10:56:58.545812 master-0 kubenswrapper[24928]: I1205 10:56:58.545738 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-d656f4996-kjkt5"] Dec 05 10:56:58.949382 master-0 kubenswrapper[24928]: I1205 10:56:58.949221 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-d656f4996-kjkt5" event={"ID":"30aa818a-f5af-4468-9920-fb1f8d1cb811","Type":"ContainerStarted","Data":"2ded5fe6d62e26efc3e9628222ed40dc36415ea6ada7ddecc6a0bb5c68225b08"} Dec 05 10:56:58.949382 master-0 kubenswrapper[24928]: I1205 10:56:58.949293 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-d656f4996-kjkt5" event={"ID":"30aa818a-f5af-4468-9920-fb1f8d1cb811","Type":"ContainerStarted","Data":"545acad0541c1a9d6068ca98ebe068536bad5bac9c64a9ce82ec7c0d3d590c94"} Dec 05 10:56:58.972172 master-0 kubenswrapper[24928]: I1205 10:56:58.972059 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-d656f4996-kjkt5" podStartSLOduration=1.9720387069999998 podStartE2EDuration="1.972038707s" podCreationTimestamp="2025-12-05 10:56:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:56:58.966716644 +0000 UTC m=+578.969910505" watchObservedRunningTime="2025-12-05 10:56:58.972038707 +0000 UTC m=+578.975232588" Dec 05 10:57:08.116785 master-0 kubenswrapper[24928]: I1205 10:57:08.116719 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:57:08.117892 master-0 kubenswrapper[24928]: I1205 10:57:08.117857 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:57:08.124281 master-0 kubenswrapper[24928]: I1205 10:57:08.124226 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:57:09.022292 master-0 kubenswrapper[24928]: I1205 10:57:09.022170 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-d656f4996-kjkt5" Dec 05 10:57:09.098255 master-0 kubenswrapper[24928]: I1205 10:57:09.098192 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-75dfb65779-zgfwv"] Dec 05 10:57:22.775271 master-0 kubenswrapper[24928]: I1205 10:57:22.775167 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["sushy-emulator/sushy-emulator-58f4c9b998-rkg8l"] Dec 05 10:57:22.777000 master-0 kubenswrapper[24928]: I1205 10:57:22.776896 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:22.780223 master-0 kubenswrapper[24928]: I1205 10:57:22.780152 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"sushy-emulator"/"sushy-emulator-config" Dec 05 10:57:22.780223 master-0 kubenswrapper[24928]: I1205 10:57:22.780206 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"sushy-emulator"/"kube-root-ca.crt" Dec 05 10:57:22.780512 master-0 kubenswrapper[24928]: I1205 10:57:22.780471 24928 reflector.go:368] Caches populated for *v1.Secret from object-"sushy-emulator"/"os-client-config" Dec 05 10:57:22.780791 master-0 kubenswrapper[24928]: I1205 10:57:22.780765 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"sushy-emulator"/"openshift-service-ca.crt" Dec 05 10:57:22.787864 master-0 kubenswrapper[24928]: I1205 10:57:22.787781 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["sushy-emulator/sushy-emulator-58f4c9b998-rkg8l"] Dec 05 10:57:22.943606 master-0 kubenswrapper[24928]: I1205 10:57:22.943527 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-client-config\" (UniqueName: \"kubernetes.io/secret/40a1ca30-6b0d-405a-9626-e23d9ab046b0-os-client-config\") pod \"sushy-emulator-58f4c9b998-rkg8l\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:22.943880 master-0 kubenswrapper[24928]: I1205 10:57:22.943667 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sushy-emulator-config\" (UniqueName: \"kubernetes.io/configmap/40a1ca30-6b0d-405a-9626-e23d9ab046b0-sushy-emulator-config\") pod \"sushy-emulator-58f4c9b998-rkg8l\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:22.944008 master-0 kubenswrapper[24928]: I1205 10:57:22.943932 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzzqh\" (UniqueName: \"kubernetes.io/projected/40a1ca30-6b0d-405a-9626-e23d9ab046b0-kube-api-access-dzzqh\") pod \"sushy-emulator-58f4c9b998-rkg8l\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:23.048530 master-0 kubenswrapper[24928]: I1205 10:57:23.048472 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sushy-emulator-config\" (UniqueName: \"kubernetes.io/configmap/40a1ca30-6b0d-405a-9626-e23d9ab046b0-sushy-emulator-config\") pod \"sushy-emulator-58f4c9b998-rkg8l\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:23.048530 master-0 kubenswrapper[24928]: I1205 10:57:23.047221 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sushy-emulator-config\" (UniqueName: \"kubernetes.io/configmap/40a1ca30-6b0d-405a-9626-e23d9ab046b0-sushy-emulator-config\") pod \"sushy-emulator-58f4c9b998-rkg8l\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:23.048860 master-0 kubenswrapper[24928]: I1205 10:57:23.048612 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzzqh\" (UniqueName: \"kubernetes.io/projected/40a1ca30-6b0d-405a-9626-e23d9ab046b0-kube-api-access-dzzqh\") pod \"sushy-emulator-58f4c9b998-rkg8l\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:23.049451 master-0 kubenswrapper[24928]: I1205 10:57:23.049398 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-client-config\" (UniqueName: \"kubernetes.io/secret/40a1ca30-6b0d-405a-9626-e23d9ab046b0-os-client-config\") pod \"sushy-emulator-58f4c9b998-rkg8l\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:23.054306 master-0 kubenswrapper[24928]: I1205 10:57:23.054252 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-client-config\" (UniqueName: \"kubernetes.io/secret/40a1ca30-6b0d-405a-9626-e23d9ab046b0-os-client-config\") pod \"sushy-emulator-58f4c9b998-rkg8l\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:23.065615 master-0 kubenswrapper[24928]: I1205 10:57:23.065538 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzzqh\" (UniqueName: \"kubernetes.io/projected/40a1ca30-6b0d-405a-9626-e23d9ab046b0-kube-api-access-dzzqh\") pod \"sushy-emulator-58f4c9b998-rkg8l\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:23.113791 master-0 kubenswrapper[24928]: I1205 10:57:23.113665 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:23.514907 master-0 kubenswrapper[24928]: I1205 10:57:23.514836 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["sushy-emulator/sushy-emulator-58f4c9b998-rkg8l"] Dec 05 10:57:23.518336 master-0 kubenswrapper[24928]: W1205 10:57:23.518234 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40a1ca30_6b0d_405a_9626_e23d9ab046b0.slice/crio-b678a2294c64d0a3cbd335132f0f5a3fb98bb375f75e1455a8b793a7036daf4b WatchSource:0}: Error finding container b678a2294c64d0a3cbd335132f0f5a3fb98bb375f75e1455a8b793a7036daf4b: Status 404 returned error can't find the container with id b678a2294c64d0a3cbd335132f0f5a3fb98bb375f75e1455a8b793a7036daf4b Dec 05 10:57:24.160076 master-0 kubenswrapper[24928]: I1205 10:57:24.159986 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" event={"ID":"40a1ca30-6b0d-405a-9626-e23d9ab046b0","Type":"ContainerStarted","Data":"b678a2294c64d0a3cbd335132f0f5a3fb98bb375f75e1455a8b793a7036daf4b"} Dec 05 10:57:26.406220 master-0 kubenswrapper[24928]: I1205 10:57:26.406143 24928 scope.go:117] "RemoveContainer" containerID="50e85593f8bade43c99e607efe092eba239c2c489a561c3d4ee019a9a7b227cb" Dec 05 10:57:30.212906 master-0 kubenswrapper[24928]: I1205 10:57:30.212831 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" event={"ID":"40a1ca30-6b0d-405a-9626-e23d9ab046b0","Type":"ContainerStarted","Data":"ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c"} Dec 05 10:57:30.248730 master-0 kubenswrapper[24928]: I1205 10:57:30.248622 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" podStartSLOduration=2.127534767 podStartE2EDuration="8.248601216s" podCreationTimestamp="2025-12-05 10:57:22 +0000 UTC" firstStartedPulling="2025-12-05 10:57:23.520734916 +0000 UTC m=+603.523928767" lastFinishedPulling="2025-12-05 10:57:29.641801355 +0000 UTC m=+609.644995216" observedRunningTime="2025-12-05 10:57:30.246274262 +0000 UTC m=+610.249468113" watchObservedRunningTime="2025-12-05 10:57:30.248601216 +0000 UTC m=+610.251795087" Dec 05 10:57:33.114671 master-0 kubenswrapper[24928]: I1205 10:57:33.114536 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:33.114671 master-0 kubenswrapper[24928]: I1205 10:57:33.114660 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:33.128992 master-0 kubenswrapper[24928]: I1205 10:57:33.128900 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:33.242949 master-0 kubenswrapper[24928]: I1205 10:57:33.242875 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 10:57:34.137020 master-0 kubenswrapper[24928]: I1205 10:57:34.136627 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-75dfb65779-zgfwv" podUID="a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" containerName="console" containerID="cri-o://d9af2b69c83d05345ecddb38e578653ae88426a35a9767f27c9efb807c0762b1" gracePeriod=15 Dec 05 10:57:35.255985 master-0 kubenswrapper[24928]: I1205 10:57:35.255941 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-75dfb65779-zgfwv_a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e/console/0.log" Dec 05 10:57:35.256660 master-0 kubenswrapper[24928]: I1205 10:57:35.256632 24928 generic.go:334] "Generic (PLEG): container finished" podID="a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" containerID="d9af2b69c83d05345ecddb38e578653ae88426a35a9767f27c9efb807c0762b1" exitCode=2 Dec 05 10:57:35.256818 master-0 kubenswrapper[24928]: I1205 10:57:35.256707 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75dfb65779-zgfwv" event={"ID":"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e","Type":"ContainerDied","Data":"d9af2b69c83d05345ecddb38e578653ae88426a35a9767f27c9efb807c0762b1"} Dec 05 10:57:35.654769 master-0 kubenswrapper[24928]: I1205 10:57:35.654649 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-75dfb65779-zgfwv_a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e/console/0.log" Dec 05 10:57:35.654769 master-0 kubenswrapper[24928]: I1205 10:57:35.654772 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:57:35.764598 master-0 kubenswrapper[24928]: I1205 10:57:35.764520 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-serving-cert\") pod \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " Dec 05 10:57:35.764598 master-0 kubenswrapper[24928]: I1205 10:57:35.764604 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-986m9\" (UniqueName: \"kubernetes.io/projected/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-kube-api-access-986m9\") pod \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " Dec 05 10:57:35.764874 master-0 kubenswrapper[24928]: I1205 10:57:35.764650 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-config\") pod \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " Dec 05 10:57:35.764874 master-0 kubenswrapper[24928]: I1205 10:57:35.764767 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-oauth-config\") pod \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " Dec 05 10:57:35.764874 master-0 kubenswrapper[24928]: I1205 10:57:35.764850 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-service-ca\") pod \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " Dec 05 10:57:35.764979 master-0 kubenswrapper[24928]: I1205 10:57:35.764877 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-trusted-ca-bundle\") pod \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " Dec 05 10:57:35.764979 master-0 kubenswrapper[24928]: I1205 10:57:35.764955 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-oauth-serving-cert\") pod \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\" (UID: \"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e\") " Dec 05 10:57:35.765390 master-0 kubenswrapper[24928]: I1205 10:57:35.765326 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-config" (OuterVolumeSpecName: "console-config") pod "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" (UID: "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:57:35.765688 master-0 kubenswrapper[24928]: I1205 10:57:35.765615 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-service-ca" (OuterVolumeSpecName: "service-ca") pod "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" (UID: "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:57:35.765736 master-0 kubenswrapper[24928]: I1205 10:57:35.765626 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" (UID: "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:57:35.765788 master-0 kubenswrapper[24928]: I1205 10:57:35.765740 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" (UID: "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:57:35.768066 master-0 kubenswrapper[24928]: I1205 10:57:35.767995 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-kube-api-access-986m9" (OuterVolumeSpecName: "kube-api-access-986m9") pod "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" (UID: "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e"). InnerVolumeSpecName "kube-api-access-986m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:57:35.768230 master-0 kubenswrapper[24928]: I1205 10:57:35.768172 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" (UID: "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:57:35.769162 master-0 kubenswrapper[24928]: I1205 10:57:35.769115 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" (UID: "a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:57:35.869544 master-0 kubenswrapper[24928]: I1205 10:57:35.866617 24928 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-oauth-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:57:35.869544 master-0 kubenswrapper[24928]: I1205 10:57:35.866678 24928 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 10:57:35.869544 master-0 kubenswrapper[24928]: I1205 10:57:35.866694 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-986m9\" (UniqueName: \"kubernetes.io/projected/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-kube-api-access-986m9\") on node \"master-0\" DevicePath \"\"" Dec 05 10:57:35.869544 master-0 kubenswrapper[24928]: I1205 10:57:35.866707 24928 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:57:35.869544 master-0 kubenswrapper[24928]: I1205 10:57:35.866719 24928 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-console-oauth-config\") on node \"master-0\" DevicePath \"\"" Dec 05 10:57:35.869544 master-0 kubenswrapper[24928]: I1205 10:57:35.866732 24928 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-service-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 10:57:35.869544 master-0 kubenswrapper[24928]: I1205 10:57:35.866745 24928 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e-trusted-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:57:36.265107 master-0 kubenswrapper[24928]: I1205 10:57:36.265059 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-75dfb65779-zgfwv_a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e/console/0.log" Dec 05 10:57:36.265881 master-0 kubenswrapper[24928]: I1205 10:57:36.265136 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75dfb65779-zgfwv" event={"ID":"a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e","Type":"ContainerDied","Data":"0bc51a3efdd9870480dec9ce3757dff7dac46efb97f0b577b15dbdbb26d34a55"} Dec 05 10:57:36.265881 master-0 kubenswrapper[24928]: I1205 10:57:36.265186 24928 scope.go:117] "RemoveContainer" containerID="d9af2b69c83d05345ecddb38e578653ae88426a35a9767f27c9efb807c0762b1" Dec 05 10:57:36.265881 master-0 kubenswrapper[24928]: I1205 10:57:36.265220 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75dfb65779-zgfwv" Dec 05 10:57:36.419515 master-0 kubenswrapper[24928]: I1205 10:57:36.419403 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-75dfb65779-zgfwv"] Dec 05 10:57:36.430483 master-0 kubenswrapper[24928]: I1205 10:57:36.430372 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-75dfb65779-zgfwv"] Dec 05 10:57:38.219463 master-0 kubenswrapper[24928]: I1205 10:57:38.219346 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" path="/var/lib/kubelet/pods/a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e/volumes" Dec 05 10:57:38.990531 master-0 kubenswrapper[24928]: I1205 10:57:38.990414 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t"] Dec 05 10:57:38.990939 master-0 kubenswrapper[24928]: E1205 10:57:38.990875 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" containerName="console" Dec 05 10:57:38.990939 master-0 kubenswrapper[24928]: I1205 10:57:38.990901 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" containerName="console" Dec 05 10:57:38.991202 master-0 kubenswrapper[24928]: I1205 10:57:38.991160 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="a92cfe85-6e5d-4307-bb0d-4bc7dfa2be6e" containerName="console" Dec 05 10:57:38.992227 master-0 kubenswrapper[24928]: I1205 10:57:38.992192 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t" Dec 05 10:57:38.994624 master-0 kubenswrapper[24928]: I1205 10:57:38.994574 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 05 10:57:38.999095 master-0 kubenswrapper[24928]: I1205 10:57:38.999038 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 05 10:57:39.001175 master-0 kubenswrapper[24928]: I1205 10:57:39.001103 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t"] Dec 05 10:57:39.015352 master-0 kubenswrapper[24928]: I1205 10:57:39.015255 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b86497c3-6d6e-4b8a-af52-46ec250f228d-nginx-conf\") pod \"networking-console-plugin-7d45bf9455-pwb9t\" (UID: \"b86497c3-6d6e-4b8a-af52-46ec250f228d\") " pod="openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t" Dec 05 10:57:39.015642 master-0 kubenswrapper[24928]: I1205 10:57:39.015515 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/b86497c3-6d6e-4b8a-af52-46ec250f228d-networking-console-plugin-cert\") pod \"networking-console-plugin-7d45bf9455-pwb9t\" (UID: \"b86497c3-6d6e-4b8a-af52-46ec250f228d\") " pod="openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t" Dec 05 10:57:39.117535 master-0 kubenswrapper[24928]: I1205 10:57:39.117435 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b86497c3-6d6e-4b8a-af52-46ec250f228d-nginx-conf\") pod \"networking-console-plugin-7d45bf9455-pwb9t\" (UID: \"b86497c3-6d6e-4b8a-af52-46ec250f228d\") " pod="openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t" Dec 05 10:57:39.117779 master-0 kubenswrapper[24928]: I1205 10:57:39.117554 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/b86497c3-6d6e-4b8a-af52-46ec250f228d-networking-console-plugin-cert\") pod \"networking-console-plugin-7d45bf9455-pwb9t\" (UID: \"b86497c3-6d6e-4b8a-af52-46ec250f228d\") " pod="openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t" Dec 05 10:57:39.118791 master-0 kubenswrapper[24928]: I1205 10:57:39.118707 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b86497c3-6d6e-4b8a-af52-46ec250f228d-nginx-conf\") pod \"networking-console-plugin-7d45bf9455-pwb9t\" (UID: \"b86497c3-6d6e-4b8a-af52-46ec250f228d\") " pod="openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t" Dec 05 10:57:39.122356 master-0 kubenswrapper[24928]: I1205 10:57:39.122288 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/b86497c3-6d6e-4b8a-af52-46ec250f228d-networking-console-plugin-cert\") pod \"networking-console-plugin-7d45bf9455-pwb9t\" (UID: \"b86497c3-6d6e-4b8a-af52-46ec250f228d\") " pod="openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t" Dec 05 10:57:39.310996 master-0 kubenswrapper[24928]: I1205 10:57:39.310876 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t" Dec 05 10:57:39.803261 master-0 kubenswrapper[24928]: I1205 10:57:39.803205 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t"] Dec 05 10:57:39.810183 master-0 kubenswrapper[24928]: W1205 10:57:39.810123 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb86497c3_6d6e_4b8a_af52_46ec250f228d.slice/crio-53f753d3b59b03c430b51cdf73a3372ef9212f50dc8589611ab63b90f55d66d9 WatchSource:0}: Error finding container 53f753d3b59b03c430b51cdf73a3372ef9212f50dc8589611ab63b90f55d66d9: Status 404 returned error can't find the container with id 53f753d3b59b03c430b51cdf73a3372ef9212f50dc8589611ab63b90f55d66d9 Dec 05 10:57:40.298163 master-0 kubenswrapper[24928]: I1205 10:57:40.298096 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t" event={"ID":"b86497c3-6d6e-4b8a-af52-46ec250f228d","Type":"ContainerStarted","Data":"53f753d3b59b03c430b51cdf73a3372ef9212f50dc8589611ab63b90f55d66d9"} Dec 05 10:57:41.315543 master-0 kubenswrapper[24928]: I1205 10:57:41.315494 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t" event={"ID":"b86497c3-6d6e-4b8a-af52-46ec250f228d","Type":"ContainerStarted","Data":"f58e1011dd11935878ca0f8990c615d0f95d3bdd7a69965d40f892b7be05501b"} Dec 05 10:57:41.340665 master-0 kubenswrapper[24928]: I1205 10:57:41.340521 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-network-console/networking-console-plugin-7d45bf9455-pwb9t" podStartSLOduration=2.060610345 podStartE2EDuration="3.340497652s" podCreationTimestamp="2025-12-05 10:57:38 +0000 UTC" firstStartedPulling="2025-12-05 10:57:39.813359339 +0000 UTC m=+619.816553200" lastFinishedPulling="2025-12-05 10:57:41.093246656 +0000 UTC m=+621.096440507" observedRunningTime="2025-12-05 10:57:41.332529587 +0000 UTC m=+621.335723448" watchObservedRunningTime="2025-12-05 10:57:41.340497652 +0000 UTC m=+621.343691503" Dec 05 10:58:04.904196 master-0 kubenswrapper[24928]: I1205 10:58:04.904125 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf"] Dec 05 10:58:04.905723 master-0 kubenswrapper[24928]: I1205 10:58:04.905695 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:04.908014 master-0 kubenswrapper[24928]: I1205 10:58:04.907929 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-2n9x6" Dec 05 10:58:04.917947 master-0 kubenswrapper[24928]: I1205 10:58:04.917884 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf"] Dec 05 10:58:04.982671 master-0 kubenswrapper[24928]: I1205 10:58:04.982583 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7g4x\" (UniqueName: \"kubernetes.io/projected/2aca0876-dc83-4922-b9c7-6528051d01fd-kube-api-access-w7g4x\") pod \"7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:04.982929 master-0 kubenswrapper[24928]: I1205 10:58:04.982754 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-util\") pod \"7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:04.982929 master-0 kubenswrapper[24928]: I1205 10:58:04.982831 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-bundle\") pod \"7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:05.083605 master-0 kubenswrapper[24928]: I1205 10:58:05.083530 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-bundle\") pod \"7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:05.083829 master-0 kubenswrapper[24928]: I1205 10:58:05.083646 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7g4x\" (UniqueName: \"kubernetes.io/projected/2aca0876-dc83-4922-b9c7-6528051d01fd-kube-api-access-w7g4x\") pod \"7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:05.083829 master-0 kubenswrapper[24928]: I1205 10:58:05.083695 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-util\") pod \"7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:05.084167 master-0 kubenswrapper[24928]: I1205 10:58:05.084140 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-util\") pod \"7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:05.084405 master-0 kubenswrapper[24928]: I1205 10:58:05.084336 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-bundle\") pod \"7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:05.099641 master-0 kubenswrapper[24928]: I1205 10:58:05.099575 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7g4x\" (UniqueName: \"kubernetes.io/projected/2aca0876-dc83-4922-b9c7-6528051d01fd-kube-api-access-w7g4x\") pod \"7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:05.223300 master-0 kubenswrapper[24928]: I1205 10:58:05.223179 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:05.610938 master-0 kubenswrapper[24928]: I1205 10:58:05.610852 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf"] Dec 05 10:58:05.619277 master-0 kubenswrapper[24928]: W1205 10:58:05.619218 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2aca0876_dc83_4922_b9c7_6528051d01fd.slice/crio-d19c6de43e379583a0b735b79465ff6dc77d711a2afc71070a38263673d2aa06 WatchSource:0}: Error finding container d19c6de43e379583a0b735b79465ff6dc77d711a2afc71070a38263673d2aa06: Status 404 returned error can't find the container with id d19c6de43e379583a0b735b79465ff6dc77d711a2afc71070a38263673d2aa06 Dec 05 10:58:06.532974 master-0 kubenswrapper[24928]: I1205 10:58:06.532884 24928 generic.go:334] "Generic (PLEG): container finished" podID="2aca0876-dc83-4922-b9c7-6528051d01fd" containerID="adba2325982a68b6d4a769c7cbbe0eb688771516c98c0375ee1a18e82ed2b081" exitCode=0 Dec 05 10:58:06.534038 master-0 kubenswrapper[24928]: I1205 10:58:06.533967 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" event={"ID":"2aca0876-dc83-4922-b9c7-6528051d01fd","Type":"ContainerDied","Data":"adba2325982a68b6d4a769c7cbbe0eb688771516c98c0375ee1a18e82ed2b081"} Dec 05 10:58:06.534246 master-0 kubenswrapper[24928]: I1205 10:58:06.534210 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" event={"ID":"2aca0876-dc83-4922-b9c7-6528051d01fd","Type":"ContainerStarted","Data":"d19c6de43e379583a0b735b79465ff6dc77d711a2afc71070a38263673d2aa06"} Dec 05 10:58:09.560758 master-0 kubenswrapper[24928]: I1205 10:58:09.560721 24928 generic.go:334] "Generic (PLEG): container finished" podID="2aca0876-dc83-4922-b9c7-6528051d01fd" containerID="6cb437307de62e67f9b5b8b3ce258cf4f3ffe7d65047598216a9a4a9ea8bc509" exitCode=0 Dec 05 10:58:09.561398 master-0 kubenswrapper[24928]: I1205 10:58:09.561360 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" event={"ID":"2aca0876-dc83-4922-b9c7-6528051d01fd","Type":"ContainerDied","Data":"6cb437307de62e67f9b5b8b3ce258cf4f3ffe7d65047598216a9a4a9ea8bc509"} Dec 05 10:58:10.584777 master-0 kubenswrapper[24928]: I1205 10:58:10.584672 24928 generic.go:334] "Generic (PLEG): container finished" podID="2aca0876-dc83-4922-b9c7-6528051d01fd" containerID="a8cae1dedde7ffd303fce76e57cc8446fb914b08193350127fcc6b25ec5707d5" exitCode=0 Dec 05 10:58:10.584777 master-0 kubenswrapper[24928]: I1205 10:58:10.584755 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" event={"ID":"2aca0876-dc83-4922-b9c7-6528051d01fd","Type":"ContainerDied","Data":"a8cae1dedde7ffd303fce76e57cc8446fb914b08193350127fcc6b25ec5707d5"} Dec 05 10:58:11.945631 master-0 kubenswrapper[24928]: I1205 10:58:11.945559 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:12.096527 master-0 kubenswrapper[24928]: I1205 10:58:12.096405 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-bundle\") pod \"2aca0876-dc83-4922-b9c7-6528051d01fd\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " Dec 05 10:58:12.096527 master-0 kubenswrapper[24928]: I1205 10:58:12.096545 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7g4x\" (UniqueName: \"kubernetes.io/projected/2aca0876-dc83-4922-b9c7-6528051d01fd-kube-api-access-w7g4x\") pod \"2aca0876-dc83-4922-b9c7-6528051d01fd\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " Dec 05 10:58:12.096905 master-0 kubenswrapper[24928]: I1205 10:58:12.096608 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-util\") pod \"2aca0876-dc83-4922-b9c7-6528051d01fd\" (UID: \"2aca0876-dc83-4922-b9c7-6528051d01fd\") " Dec 05 10:58:12.097290 master-0 kubenswrapper[24928]: I1205 10:58:12.097224 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-bundle" (OuterVolumeSpecName: "bundle") pod "2aca0876-dc83-4922-b9c7-6528051d01fd" (UID: "2aca0876-dc83-4922-b9c7-6528051d01fd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:58:12.099383 master-0 kubenswrapper[24928]: I1205 10:58:12.099323 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aca0876-dc83-4922-b9c7-6528051d01fd-kube-api-access-w7g4x" (OuterVolumeSpecName: "kube-api-access-w7g4x") pod "2aca0876-dc83-4922-b9c7-6528051d01fd" (UID: "2aca0876-dc83-4922-b9c7-6528051d01fd"). InnerVolumeSpecName "kube-api-access-w7g4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:58:12.111808 master-0 kubenswrapper[24928]: I1205 10:58:12.111735 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-util" (OuterVolumeSpecName: "util") pod "2aca0876-dc83-4922-b9c7-6528051d01fd" (UID: "2aca0876-dc83-4922-b9c7-6528051d01fd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:58:12.198136 master-0 kubenswrapper[24928]: I1205 10:58:12.198066 24928 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-util\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:12.198136 master-0 kubenswrapper[24928]: I1205 10:58:12.198124 24928 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2aca0876-dc83-4922-b9c7-6528051d01fd-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:12.198136 master-0 kubenswrapper[24928]: I1205 10:58:12.198146 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7g4x\" (UniqueName: \"kubernetes.io/projected/2aca0876-dc83-4922-b9c7-6528051d01fd-kube-api-access-w7g4x\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:12.602094 master-0 kubenswrapper[24928]: I1205 10:58:12.602015 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" event={"ID":"2aca0876-dc83-4922-b9c7-6528051d01fd","Type":"ContainerDied","Data":"d19c6de43e379583a0b735b79465ff6dc77d711a2afc71070a38263673d2aa06"} Dec 05 10:58:12.602094 master-0 kubenswrapper[24928]: I1205 10:58:12.602081 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d19c6de43e379583a0b735b79465ff6dc77d711a2afc71070a38263673d2aa06" Dec 05 10:58:12.602457 master-0 kubenswrapper[24928]: I1205 10:58:12.602106 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7f6062bfcf66f08711c4d599873349559e66916847a22b4b74a32f97d48ztzf" Dec 05 10:58:18.267221 master-0 kubenswrapper[24928]: I1205 10:58:18.267144 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-storage/lvms-operator-d7bbfbfb7-js4fd"] Dec 05 10:58:18.267847 master-0 kubenswrapper[24928]: E1205 10:58:18.267503 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aca0876-dc83-4922-b9c7-6528051d01fd" containerName="util" Dec 05 10:58:18.267847 master-0 kubenswrapper[24928]: I1205 10:58:18.267522 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aca0876-dc83-4922-b9c7-6528051d01fd" containerName="util" Dec 05 10:58:18.267847 master-0 kubenswrapper[24928]: E1205 10:58:18.267553 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aca0876-dc83-4922-b9c7-6528051d01fd" containerName="extract" Dec 05 10:58:18.267847 master-0 kubenswrapper[24928]: I1205 10:58:18.267564 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aca0876-dc83-4922-b9c7-6528051d01fd" containerName="extract" Dec 05 10:58:18.267847 master-0 kubenswrapper[24928]: E1205 10:58:18.267582 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aca0876-dc83-4922-b9c7-6528051d01fd" containerName="pull" Dec 05 10:58:18.267847 master-0 kubenswrapper[24928]: I1205 10:58:18.267590 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aca0876-dc83-4922-b9c7-6528051d01fd" containerName="pull" Dec 05 10:58:18.267847 master-0 kubenswrapper[24928]: I1205 10:58:18.267803 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aca0876-dc83-4922-b9c7-6528051d01fd" containerName="extract" Dec 05 10:58:18.268382 master-0 kubenswrapper[24928]: I1205 10:58:18.268354 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.278766 master-0 kubenswrapper[24928]: I1205 10:58:18.278732 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-storage"/"lvms-operator-metrics-cert" Dec 05 10:58:18.280275 master-0 kubenswrapper[24928]: I1205 10:58:18.280232 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-storage"/"openshift-service-ca.crt" Dec 05 10:58:18.281268 master-0 kubenswrapper[24928]: I1205 10:58:18.281217 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-storage"/"kube-root-ca.crt" Dec 05 10:58:18.281341 master-0 kubenswrapper[24928]: I1205 10:58:18.281294 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-storage"/"lvms-operator-service-cert" Dec 05 10:58:18.281603 master-0 kubenswrapper[24928]: I1205 10:58:18.281574 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-storage"/"lvms-operator-webhook-server-cert" Dec 05 10:58:18.304757 master-0 kubenswrapper[24928]: I1205 10:58:18.304675 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-storage/lvms-operator-d7bbfbfb7-js4fd"] Dec 05 10:58:18.392739 master-0 kubenswrapper[24928]: I1205 10:58:18.392669 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7b9ea4be-adf5-4e90-815f-21e94967d8fd-apiservice-cert\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.392739 master-0 kubenswrapper[24928]: I1205 10:58:18.392749 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dm6f\" (UniqueName: \"kubernetes.io/projected/7b9ea4be-adf5-4e90-815f-21e94967d8fd-kube-api-access-4dm6f\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.393005 master-0 kubenswrapper[24928]: I1205 10:58:18.392837 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/7b9ea4be-adf5-4e90-815f-21e94967d8fd-socket-dir\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.393005 master-0 kubenswrapper[24928]: I1205 10:58:18.392869 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/7b9ea4be-adf5-4e90-815f-21e94967d8fd-metrics-cert\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.393005 master-0 kubenswrapper[24928]: I1205 10:58:18.392888 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7b9ea4be-adf5-4e90-815f-21e94967d8fd-webhook-cert\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.494767 master-0 kubenswrapper[24928]: I1205 10:58:18.494687 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/7b9ea4be-adf5-4e90-815f-21e94967d8fd-socket-dir\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.494767 master-0 kubenswrapper[24928]: I1205 10:58:18.494761 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/7b9ea4be-adf5-4e90-815f-21e94967d8fd-metrics-cert\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.495069 master-0 kubenswrapper[24928]: I1205 10:58:18.494791 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7b9ea4be-adf5-4e90-815f-21e94967d8fd-webhook-cert\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.495069 master-0 kubenswrapper[24928]: I1205 10:58:18.494837 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7b9ea4be-adf5-4e90-815f-21e94967d8fd-apiservice-cert\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.495069 master-0 kubenswrapper[24928]: I1205 10:58:18.494884 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dm6f\" (UniqueName: \"kubernetes.io/projected/7b9ea4be-adf5-4e90-815f-21e94967d8fd-kube-api-access-4dm6f\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.495361 master-0 kubenswrapper[24928]: I1205 10:58:18.495323 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/7b9ea4be-adf5-4e90-815f-21e94967d8fd-socket-dir\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.498255 master-0 kubenswrapper[24928]: I1205 10:58:18.498197 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/7b9ea4be-adf5-4e90-815f-21e94967d8fd-metrics-cert\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.498683 master-0 kubenswrapper[24928]: I1205 10:58:18.498637 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7b9ea4be-adf5-4e90-815f-21e94967d8fd-webhook-cert\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.499286 master-0 kubenswrapper[24928]: I1205 10:58:18.499266 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7b9ea4be-adf5-4e90-815f-21e94967d8fd-apiservice-cert\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.513313 master-0 kubenswrapper[24928]: I1205 10:58:18.513260 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dm6f\" (UniqueName: \"kubernetes.io/projected/7b9ea4be-adf5-4e90-815f-21e94967d8fd-kube-api-access-4dm6f\") pod \"lvms-operator-d7bbfbfb7-js4fd\" (UID: \"7b9ea4be-adf5-4e90-815f-21e94967d8fd\") " pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:18.582200 master-0 kubenswrapper[24928]: I1205 10:58:18.582128 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:19.076764 master-0 kubenswrapper[24928]: I1205 10:58:19.075287 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-storage/lvms-operator-d7bbfbfb7-js4fd"] Dec 05 10:58:19.663661 master-0 kubenswrapper[24928]: I1205 10:58:19.663593 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" event={"ID":"7b9ea4be-adf5-4e90-815f-21e94967d8fd","Type":"ContainerStarted","Data":"d17d0d85b080dd15d36f6ef413df6d7766195af20f52564d82f7d63b4c656213"} Dec 05 10:58:23.697257 master-0 kubenswrapper[24928]: I1205 10:58:23.697104 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" event={"ID":"7b9ea4be-adf5-4e90-815f-21e94967d8fd","Type":"ContainerStarted","Data":"ced18aef00e0711375a28999ac710c0b4f7f75ad44d30d4a61131760dc0a72ce"} Dec 05 10:58:23.697917 master-0 kubenswrapper[24928]: I1205 10:58:23.697304 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:23.720858 master-0 kubenswrapper[24928]: I1205 10:58:23.720753 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" podStartSLOduration=1.395603612 podStartE2EDuration="5.720734289s" podCreationTimestamp="2025-12-05 10:58:18 +0000 UTC" firstStartedPulling="2025-12-05 10:58:19.084733017 +0000 UTC m=+659.087926868" lastFinishedPulling="2025-12-05 10:58:23.409863694 +0000 UTC m=+663.413057545" observedRunningTime="2025-12-05 10:58:23.716754622 +0000 UTC m=+663.719948533" watchObservedRunningTime="2025-12-05 10:58:23.720734289 +0000 UTC m=+663.723928130" Dec 05 10:58:24.708152 master-0 kubenswrapper[24928]: I1205 10:58:24.708061 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-storage/lvms-operator-d7bbfbfb7-js4fd" Dec 05 10:58:28.696266 master-0 kubenswrapper[24928]: I1205 10:58:28.696192 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h"] Dec 05 10:58:28.697786 master-0 kubenswrapper[24928]: I1205 10:58:28.697753 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:28.700211 master-0 kubenswrapper[24928]: I1205 10:58:28.700170 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-2n9x6" Dec 05 10:58:28.707541 master-0 kubenswrapper[24928]: I1205 10:58:28.707477 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h"] Dec 05 10:58:28.721212 master-0 kubenswrapper[24928]: I1205 10:58:28.721146 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbgqj\" (UniqueName: \"kubernetes.io/projected/e285e7ab-e929-426a-9d59-90a1382339e2-kube-api-access-gbgqj\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:28.721554 master-0 kubenswrapper[24928]: I1205 10:58:28.721223 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:28.721554 master-0 kubenswrapper[24928]: I1205 10:58:28.721272 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:28.822021 master-0 kubenswrapper[24928]: I1205 10:58:28.821940 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:28.822021 master-0 kubenswrapper[24928]: I1205 10:58:28.822038 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:28.822438 master-0 kubenswrapper[24928]: I1205 10:58:28.822093 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbgqj\" (UniqueName: \"kubernetes.io/projected/e285e7ab-e929-426a-9d59-90a1382339e2-kube-api-access-gbgqj\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:28.822438 master-0 kubenswrapper[24928]: I1205 10:58:28.822387 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:28.822594 master-0 kubenswrapper[24928]: I1205 10:58:28.822546 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:28.846401 master-0 kubenswrapper[24928]: I1205 10:58:28.846326 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbgqj\" (UniqueName: \"kubernetes.io/projected/e285e7ab-e929-426a-9d59-90a1382339e2-kube-api-access-gbgqj\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:29.021576 master-0 kubenswrapper[24928]: I1205 10:58:29.021385 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:29.464009 master-0 kubenswrapper[24928]: I1205 10:58:29.463923 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h"] Dec 05 10:58:29.464697 master-0 kubenswrapper[24928]: W1205 10:58:29.464632 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode285e7ab_e929_426a_9d59_90a1382339e2.slice/crio-e4e01dbd811098dc30fc10c4b777a81f87962ccafed25d20a81251f94c82b337 WatchSource:0}: Error finding container e4e01dbd811098dc30fc10c4b777a81f87962ccafed25d20a81251f94c82b337: Status 404 returned error can't find the container with id e4e01dbd811098dc30fc10c4b777a81f87962ccafed25d20a81251f94c82b337 Dec 05 10:58:29.578877 master-0 kubenswrapper[24928]: I1205 10:58:29.578781 24928 scope.go:117] "RemoveContainer" containerID="b27b1cff74cf1fa7645c9d670ba8742571627f768b820d46bb086bfbeb3ea29b" Dec 05 10:58:29.817918 master-0 kubenswrapper[24928]: I1205 10:58:29.817800 24928 generic.go:334] "Generic (PLEG): container finished" podID="e285e7ab-e929-426a-9d59-90a1382339e2" containerID="b3b9c8b94b389902f75d547e53fecedf54fd580c402468214a6b02ff05b58300" exitCode=0 Dec 05 10:58:29.817918 master-0 kubenswrapper[24928]: I1205 10:58:29.817859 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" event={"ID":"e285e7ab-e929-426a-9d59-90a1382339e2","Type":"ContainerDied","Data":"b3b9c8b94b389902f75d547e53fecedf54fd580c402468214a6b02ff05b58300"} Dec 05 10:58:29.817918 master-0 kubenswrapper[24928]: I1205 10:58:29.817889 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" event={"ID":"e285e7ab-e929-426a-9d59-90a1382339e2","Type":"ContainerStarted","Data":"e4e01dbd811098dc30fc10c4b777a81f87962ccafed25d20a81251f94c82b337"} Dec 05 10:58:30.923451 master-0 kubenswrapper[24928]: I1205 10:58:30.921388 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5"] Dec 05 10:58:30.923451 master-0 kubenswrapper[24928]: I1205 10:58:30.922862 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:30.936539 master-0 kubenswrapper[24928]: I1205 10:58:30.936209 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5"] Dec 05 10:58:31.075119 master-0 kubenswrapper[24928]: I1205 10:58:31.075057 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:31.075325 master-0 kubenswrapper[24928]: I1205 10:58:31.075168 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnc4l\" (UniqueName: \"kubernetes.io/projected/6d2ede2f-19b3-4665-b465-0651ab17b057-kube-api-access-mnc4l\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:31.075325 master-0 kubenswrapper[24928]: I1205 10:58:31.075214 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:31.177117 master-0 kubenswrapper[24928]: I1205 10:58:31.176965 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnc4l\" (UniqueName: \"kubernetes.io/projected/6d2ede2f-19b3-4665-b465-0651ab17b057-kube-api-access-mnc4l\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:31.177117 master-0 kubenswrapper[24928]: I1205 10:58:31.177100 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:31.177368 master-0 kubenswrapper[24928]: I1205 10:58:31.177337 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:31.177797 master-0 kubenswrapper[24928]: I1205 10:58:31.177752 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:31.177868 master-0 kubenswrapper[24928]: I1205 10:58:31.177832 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:31.204477 master-0 kubenswrapper[24928]: I1205 10:58:31.204415 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnc4l\" (UniqueName: \"kubernetes.io/projected/6d2ede2f-19b3-4665-b465-0651ab17b057-kube-api-access-mnc4l\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:31.266204 master-0 kubenswrapper[24928]: I1205 10:58:31.266104 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:31.719734 master-0 kubenswrapper[24928]: I1205 10:58:31.719563 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x"] Dec 05 10:58:31.721268 master-0 kubenswrapper[24928]: I1205 10:58:31.721242 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:31.736723 master-0 kubenswrapper[24928]: I1205 10:58:31.736442 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x"] Dec 05 10:58:31.743876 master-0 kubenswrapper[24928]: I1205 10:58:31.743807 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5"] Dec 05 10:58:31.887780 master-0 kubenswrapper[24928]: I1205 10:58:31.887693 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:31.887780 master-0 kubenswrapper[24928]: I1205 10:58:31.887775 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wwt5\" (UniqueName: \"kubernetes.io/projected/404e6a6f-7745-4aef-b053-8a483c6baee9-kube-api-access-6wwt5\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:31.888062 master-0 kubenswrapper[24928]: I1205 10:58:31.887820 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:31.990042 master-0 kubenswrapper[24928]: I1205 10:58:31.989780 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:31.990042 master-0 kubenswrapper[24928]: I1205 10:58:31.989930 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wwt5\" (UniqueName: \"kubernetes.io/projected/404e6a6f-7745-4aef-b053-8a483c6baee9-kube-api-access-6wwt5\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:31.990042 master-0 kubenswrapper[24928]: I1205 10:58:31.990016 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:31.990996 master-0 kubenswrapper[24928]: I1205 10:58:31.990852 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:31.991205 master-0 kubenswrapper[24928]: I1205 10:58:31.991125 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:32.008227 master-0 kubenswrapper[24928]: I1205 10:58:32.008149 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wwt5\" (UniqueName: \"kubernetes.io/projected/404e6a6f-7745-4aef-b053-8a483c6baee9-kube-api-access-6wwt5\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:32.050685 master-0 kubenswrapper[24928]: I1205 10:58:32.050625 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:32.647963 master-0 kubenswrapper[24928]: W1205 10:58:32.647863 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d2ede2f_19b3_4665_b465_0651ab17b057.slice/crio-758949d9678c185f9d998b850a57f3c84961537f4ecf75291fe42ed15ad4b743 WatchSource:0}: Error finding container 758949d9678c185f9d998b850a57f3c84961537f4ecf75291fe42ed15ad4b743: Status 404 returned error can't find the container with id 758949d9678c185f9d998b850a57f3c84961537f4ecf75291fe42ed15ad4b743 Dec 05 10:58:32.845441 master-0 kubenswrapper[24928]: I1205 10:58:32.845370 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" event={"ID":"6d2ede2f-19b3-4665-b465-0651ab17b057","Type":"ContainerStarted","Data":"758949d9678c185f9d998b850a57f3c84961537f4ecf75291fe42ed15ad4b743"} Dec 05 10:58:33.105772 master-0 kubenswrapper[24928]: I1205 10:58:33.105698 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x"] Dec 05 10:58:33.138603 master-0 kubenswrapper[24928]: W1205 10:58:33.138543 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod404e6a6f_7745_4aef_b053_8a483c6baee9.slice/crio-a8db2a9f984afb0b9dce3258e627287b22b3ce25f3cacd4f76acfece67e65c47 WatchSource:0}: Error finding container a8db2a9f984afb0b9dce3258e627287b22b3ce25f3cacd4f76acfece67e65c47: Status 404 returned error can't find the container with id a8db2a9f984afb0b9dce3258e627287b22b3ce25f3cacd4f76acfece67e65c47 Dec 05 10:58:33.857005 master-0 kubenswrapper[24928]: I1205 10:58:33.856861 24928 generic.go:334] "Generic (PLEG): container finished" podID="404e6a6f-7745-4aef-b053-8a483c6baee9" containerID="2344585191bd086fa9310489d6a97e2f2798d663915230a7781f62376b9fba25" exitCode=0 Dec 05 10:58:33.857005 master-0 kubenswrapper[24928]: I1205 10:58:33.856936 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" event={"ID":"404e6a6f-7745-4aef-b053-8a483c6baee9","Type":"ContainerDied","Data":"2344585191bd086fa9310489d6a97e2f2798d663915230a7781f62376b9fba25"} Dec 05 10:58:33.857005 master-0 kubenswrapper[24928]: I1205 10:58:33.857016 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" event={"ID":"404e6a6f-7745-4aef-b053-8a483c6baee9","Type":"ContainerStarted","Data":"a8db2a9f984afb0b9dce3258e627287b22b3ce25f3cacd4f76acfece67e65c47"} Dec 05 10:58:33.861413 master-0 kubenswrapper[24928]: I1205 10:58:33.861202 24928 generic.go:334] "Generic (PLEG): container finished" podID="e285e7ab-e929-426a-9d59-90a1382339e2" containerID="a13849aed6c0be2fa7dc34eaaa1b89050a20c27776e250ef153392238aba7f30" exitCode=0 Dec 05 10:58:33.861413 master-0 kubenswrapper[24928]: I1205 10:58:33.861266 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" event={"ID":"e285e7ab-e929-426a-9d59-90a1382339e2","Type":"ContainerDied","Data":"a13849aed6c0be2fa7dc34eaaa1b89050a20c27776e250ef153392238aba7f30"} Dec 05 10:58:33.867266 master-0 kubenswrapper[24928]: I1205 10:58:33.867170 24928 generic.go:334] "Generic (PLEG): container finished" podID="6d2ede2f-19b3-4665-b465-0651ab17b057" containerID="edecda127121fd04cb77426bed92b158709470d3347da3bb08a80dbdfe6a6998" exitCode=0 Dec 05 10:58:33.867399 master-0 kubenswrapper[24928]: I1205 10:58:33.867254 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" event={"ID":"6d2ede2f-19b3-4665-b465-0651ab17b057","Type":"ContainerDied","Data":"edecda127121fd04cb77426bed92b158709470d3347da3bb08a80dbdfe6a6998"} Dec 05 10:58:34.876998 master-0 kubenswrapper[24928]: I1205 10:58:34.876945 24928 generic.go:334] "Generic (PLEG): container finished" podID="e285e7ab-e929-426a-9d59-90a1382339e2" containerID="2fc47244be22e9f743c50a0faae634fc738bee47fdd4369c025efa16ae523acc" exitCode=0 Dec 05 10:58:34.893780 master-0 kubenswrapper[24928]: I1205 10:58:34.877001 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" event={"ID":"e285e7ab-e929-426a-9d59-90a1382339e2","Type":"ContainerDied","Data":"2fc47244be22e9f743c50a0faae634fc738bee47fdd4369c025efa16ae523acc"} Dec 05 10:58:35.888756 master-0 kubenswrapper[24928]: I1205 10:58:35.888632 24928 generic.go:334] "Generic (PLEG): container finished" podID="404e6a6f-7745-4aef-b053-8a483c6baee9" containerID="c2956de22810239689e43db74b4ad8d3a80e525e378e8c83527d4f782d0b9221" exitCode=0 Dec 05 10:58:35.889293 master-0 kubenswrapper[24928]: I1205 10:58:35.888735 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" event={"ID":"404e6a6f-7745-4aef-b053-8a483c6baee9","Type":"ContainerDied","Data":"c2956de22810239689e43db74b4ad8d3a80e525e378e8c83527d4f782d0b9221"} Dec 05 10:58:35.893778 master-0 kubenswrapper[24928]: I1205 10:58:35.893747 24928 generic.go:334] "Generic (PLEG): container finished" podID="6d2ede2f-19b3-4665-b465-0651ab17b057" containerID="7795104dc3127d4148d1daff564fd79065b0eb98b15d638f0acb228fdceac9a4" exitCode=0 Dec 05 10:58:35.893957 master-0 kubenswrapper[24928]: I1205 10:58:35.893784 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" event={"ID":"6d2ede2f-19b3-4665-b465-0651ab17b057","Type":"ContainerDied","Data":"7795104dc3127d4148d1daff564fd79065b0eb98b15d638f0acb228fdceac9a4"} Dec 05 10:58:36.326791 master-0 kubenswrapper[24928]: I1205 10:58:36.326735 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:36.466188 master-0 kubenswrapper[24928]: I1205 10:58:36.466015 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-util\") pod \"e285e7ab-e929-426a-9d59-90a1382339e2\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " Dec 05 10:58:36.466188 master-0 kubenswrapper[24928]: I1205 10:58:36.466126 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-bundle\") pod \"e285e7ab-e929-426a-9d59-90a1382339e2\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " Dec 05 10:58:36.466526 master-0 kubenswrapper[24928]: I1205 10:58:36.466294 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbgqj\" (UniqueName: \"kubernetes.io/projected/e285e7ab-e929-426a-9d59-90a1382339e2-kube-api-access-gbgqj\") pod \"e285e7ab-e929-426a-9d59-90a1382339e2\" (UID: \"e285e7ab-e929-426a-9d59-90a1382339e2\") " Dec 05 10:58:36.467927 master-0 kubenswrapper[24928]: I1205 10:58:36.467816 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-bundle" (OuterVolumeSpecName: "bundle") pod "e285e7ab-e929-426a-9d59-90a1382339e2" (UID: "e285e7ab-e929-426a-9d59-90a1382339e2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:58:36.471206 master-0 kubenswrapper[24928]: I1205 10:58:36.471136 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e285e7ab-e929-426a-9d59-90a1382339e2-kube-api-access-gbgqj" (OuterVolumeSpecName: "kube-api-access-gbgqj") pod "e285e7ab-e929-426a-9d59-90a1382339e2" (UID: "e285e7ab-e929-426a-9d59-90a1382339e2"). InnerVolumeSpecName "kube-api-access-gbgqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:58:36.477086 master-0 kubenswrapper[24928]: I1205 10:58:36.477028 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-util" (OuterVolumeSpecName: "util") pod "e285e7ab-e929-426a-9d59-90a1382339e2" (UID: "e285e7ab-e929-426a-9d59-90a1382339e2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:58:36.568138 master-0 kubenswrapper[24928]: I1205 10:58:36.568054 24928 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-util\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:36.568138 master-0 kubenswrapper[24928]: I1205 10:58:36.568096 24928 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e285e7ab-e929-426a-9d59-90a1382339e2-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:36.568138 master-0 kubenswrapper[24928]: I1205 10:58:36.568113 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbgqj\" (UniqueName: \"kubernetes.io/projected/e285e7ab-e929-426a-9d59-90a1382339e2-kube-api-access-gbgqj\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:36.903640 master-0 kubenswrapper[24928]: I1205 10:58:36.903533 24928 generic.go:334] "Generic (PLEG): container finished" podID="404e6a6f-7745-4aef-b053-8a483c6baee9" containerID="8ae27fa7210beb3dc9274cacf9123e2c8603fe3c66039de51ee3e86b18e9c701" exitCode=0 Dec 05 10:58:36.904619 master-0 kubenswrapper[24928]: I1205 10:58:36.903644 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" event={"ID":"404e6a6f-7745-4aef-b053-8a483c6baee9","Type":"ContainerDied","Data":"8ae27fa7210beb3dc9274cacf9123e2c8603fe3c66039de51ee3e86b18e9c701"} Dec 05 10:58:36.908641 master-0 kubenswrapper[24928]: I1205 10:58:36.908577 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" event={"ID":"e285e7ab-e929-426a-9d59-90a1382339e2","Type":"ContainerDied","Data":"e4e01dbd811098dc30fc10c4b777a81f87962ccafed25d20a81251f94c82b337"} Dec 05 10:58:36.908641 master-0 kubenswrapper[24928]: I1205 10:58:36.908621 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4e01dbd811098dc30fc10c4b777a81f87962ccafed25d20a81251f94c82b337" Dec 05 10:58:36.908641 master-0 kubenswrapper[24928]: I1205 10:58:36.908616 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931a7c59h" Dec 05 10:58:36.911809 master-0 kubenswrapper[24928]: I1205 10:58:36.911738 24928 generic.go:334] "Generic (PLEG): container finished" podID="6d2ede2f-19b3-4665-b465-0651ab17b057" containerID="fc94b6428e806c6c79ed866f1ff5d1f2219a9109b7c28fd290dbdaa2eb0a1482" exitCode=0 Dec 05 10:58:36.911809 master-0 kubenswrapper[24928]: I1205 10:58:36.911794 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" event={"ID":"6d2ede2f-19b3-4665-b465-0651ab17b057","Type":"ContainerDied","Data":"fc94b6428e806c6c79ed866f1ff5d1f2219a9109b7c28fd290dbdaa2eb0a1482"} Dec 05 10:58:38.378473 master-0 kubenswrapper[24928]: I1205 10:58:38.378399 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:38.383090 master-0 kubenswrapper[24928]: I1205 10:58:38.383050 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:38.412720 master-0 kubenswrapper[24928]: I1205 10:58:38.412656 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-bundle\") pod \"404e6a6f-7745-4aef-b053-8a483c6baee9\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " Dec 05 10:58:38.413034 master-0 kubenswrapper[24928]: I1205 10:58:38.412750 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-util\") pod \"404e6a6f-7745-4aef-b053-8a483c6baee9\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " Dec 05 10:58:38.413034 master-0 kubenswrapper[24928]: I1205 10:58:38.412789 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-bundle\") pod \"6d2ede2f-19b3-4665-b465-0651ab17b057\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " Dec 05 10:58:38.413034 master-0 kubenswrapper[24928]: I1205 10:58:38.412871 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wwt5\" (UniqueName: \"kubernetes.io/projected/404e6a6f-7745-4aef-b053-8a483c6baee9-kube-api-access-6wwt5\") pod \"404e6a6f-7745-4aef-b053-8a483c6baee9\" (UID: \"404e6a6f-7745-4aef-b053-8a483c6baee9\") " Dec 05 10:58:38.413034 master-0 kubenswrapper[24928]: I1205 10:58:38.412902 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnc4l\" (UniqueName: \"kubernetes.io/projected/6d2ede2f-19b3-4665-b465-0651ab17b057-kube-api-access-mnc4l\") pod \"6d2ede2f-19b3-4665-b465-0651ab17b057\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " Dec 05 10:58:38.413034 master-0 kubenswrapper[24928]: I1205 10:58:38.412954 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-util\") pod \"6d2ede2f-19b3-4665-b465-0651ab17b057\" (UID: \"6d2ede2f-19b3-4665-b465-0651ab17b057\") " Dec 05 10:58:38.413252 master-0 kubenswrapper[24928]: I1205 10:58:38.413212 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-bundle" (OuterVolumeSpecName: "bundle") pod "404e6a6f-7745-4aef-b053-8a483c6baee9" (UID: "404e6a6f-7745-4aef-b053-8a483c6baee9"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:58:38.414642 master-0 kubenswrapper[24928]: I1205 10:58:38.414617 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-bundle" (OuterVolumeSpecName: "bundle") pod "6d2ede2f-19b3-4665-b465-0651ab17b057" (UID: "6d2ede2f-19b3-4665-b465-0651ab17b057"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:58:38.414927 master-0 kubenswrapper[24928]: I1205 10:58:38.414882 24928 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:38.414927 master-0 kubenswrapper[24928]: I1205 10:58:38.414919 24928 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:38.417023 master-0 kubenswrapper[24928]: I1205 10:58:38.416986 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d2ede2f-19b3-4665-b465-0651ab17b057-kube-api-access-mnc4l" (OuterVolumeSpecName: "kube-api-access-mnc4l") pod "6d2ede2f-19b3-4665-b465-0651ab17b057" (UID: "6d2ede2f-19b3-4665-b465-0651ab17b057"). InnerVolumeSpecName "kube-api-access-mnc4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:58:38.420207 master-0 kubenswrapper[24928]: I1205 10:58:38.420157 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/404e6a6f-7745-4aef-b053-8a483c6baee9-kube-api-access-6wwt5" (OuterVolumeSpecName: "kube-api-access-6wwt5") pod "404e6a6f-7745-4aef-b053-8a483c6baee9" (UID: "404e6a6f-7745-4aef-b053-8a483c6baee9"). InnerVolumeSpecName "kube-api-access-6wwt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:58:38.425036 master-0 kubenswrapper[24928]: I1205 10:58:38.424977 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-util" (OuterVolumeSpecName: "util") pod "6d2ede2f-19b3-4665-b465-0651ab17b057" (UID: "6d2ede2f-19b3-4665-b465-0651ab17b057"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:58:38.436778 master-0 kubenswrapper[24928]: I1205 10:58:38.436373 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-util" (OuterVolumeSpecName: "util") pod "404e6a6f-7745-4aef-b053-8a483c6baee9" (UID: "404e6a6f-7745-4aef-b053-8a483c6baee9"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:58:38.516618 master-0 kubenswrapper[24928]: I1205 10:58:38.516529 24928 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6d2ede2f-19b3-4665-b465-0651ab17b057-util\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:38.516618 master-0 kubenswrapper[24928]: I1205 10:58:38.516595 24928 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/404e6a6f-7745-4aef-b053-8a483c6baee9-util\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:38.516618 master-0 kubenswrapper[24928]: I1205 10:58:38.516612 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wwt5\" (UniqueName: \"kubernetes.io/projected/404e6a6f-7745-4aef-b053-8a483c6baee9-kube-api-access-6wwt5\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:38.516618 master-0 kubenswrapper[24928]: I1205 10:58:38.516630 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnc4l\" (UniqueName: \"kubernetes.io/projected/6d2ede2f-19b3-4665-b465-0651ab17b057-kube-api-access-mnc4l\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:38.935683 master-0 kubenswrapper[24928]: I1205 10:58:38.935618 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" event={"ID":"6d2ede2f-19b3-4665-b465-0651ab17b057","Type":"ContainerDied","Data":"758949d9678c185f9d998b850a57f3c84961537f4ecf75291fe42ed15ad4b743"} Dec 05 10:58:38.935683 master-0 kubenswrapper[24928]: I1205 10:58:38.935678 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="758949d9678c185f9d998b850a57f3c84961537f4ecf75291fe42ed15ad4b743" Dec 05 10:58:38.935683 master-0 kubenswrapper[24928]: I1205 10:58:38.935643 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83996n5" Dec 05 10:58:38.939052 master-0 kubenswrapper[24928]: I1205 10:58:38.939016 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" event={"ID":"404e6a6f-7745-4aef-b053-8a483c6baee9","Type":"ContainerDied","Data":"a8db2a9f984afb0b9dce3258e627287b22b3ce25f3cacd4f76acfece67e65c47"} Dec 05 10:58:38.939052 master-0 kubenswrapper[24928]: I1205 10:58:38.939049 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8db2a9f984afb0b9dce3258e627287b22b3ce25f3cacd4f76acfece67e65c47" Dec 05 10:58:38.939326 master-0 kubenswrapper[24928]: I1205 10:58:38.939295 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f5z79x" Dec 05 10:58:39.337339 master-0 kubenswrapper[24928]: I1205 10:58:39.337261 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8"] Dec 05 10:58:39.337805 master-0 kubenswrapper[24928]: E1205 10:58:39.337778 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="404e6a6f-7745-4aef-b053-8a483c6baee9" containerName="extract" Dec 05 10:58:39.337805 master-0 kubenswrapper[24928]: I1205 10:58:39.337797 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="404e6a6f-7745-4aef-b053-8a483c6baee9" containerName="extract" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: E1205 10:58:39.337821 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e285e7ab-e929-426a-9d59-90a1382339e2" containerName="extract" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: I1205 10:58:39.337828 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e285e7ab-e929-426a-9d59-90a1382339e2" containerName="extract" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: E1205 10:58:39.337843 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2ede2f-19b3-4665-b465-0651ab17b057" containerName="pull" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: I1205 10:58:39.337850 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2ede2f-19b3-4665-b465-0651ab17b057" containerName="pull" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: E1205 10:58:39.337875 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2ede2f-19b3-4665-b465-0651ab17b057" containerName="extract" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: I1205 10:58:39.337882 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2ede2f-19b3-4665-b465-0651ab17b057" containerName="extract" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: E1205 10:58:39.337898 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e285e7ab-e929-426a-9d59-90a1382339e2" containerName="util" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: I1205 10:58:39.337906 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e285e7ab-e929-426a-9d59-90a1382339e2" containerName="util" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: E1205 10:58:39.337918 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="404e6a6f-7745-4aef-b053-8a483c6baee9" containerName="util" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: I1205 10:58:39.337925 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="404e6a6f-7745-4aef-b053-8a483c6baee9" containerName="util" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: E1205 10:58:39.337943 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d2ede2f-19b3-4665-b465-0651ab17b057" containerName="util" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: I1205 10:58:39.337949 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d2ede2f-19b3-4665-b465-0651ab17b057" containerName="util" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: E1205 10:58:39.337965 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="404e6a6f-7745-4aef-b053-8a483c6baee9" containerName="pull" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: I1205 10:58:39.337971 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="404e6a6f-7745-4aef-b053-8a483c6baee9" containerName="pull" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: E1205 10:58:39.337990 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e285e7ab-e929-426a-9d59-90a1382339e2" containerName="pull" Dec 05 10:58:39.338010 master-0 kubenswrapper[24928]: I1205 10:58:39.337996 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e285e7ab-e929-426a-9d59-90a1382339e2" containerName="pull" Dec 05 10:58:39.339254 master-0 kubenswrapper[24928]: I1205 10:58:39.338276 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="404e6a6f-7745-4aef-b053-8a483c6baee9" containerName="extract" Dec 05 10:58:39.339254 master-0 kubenswrapper[24928]: I1205 10:58:39.338295 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d2ede2f-19b3-4665-b465-0651ab17b057" containerName="extract" Dec 05 10:58:39.339254 master-0 kubenswrapper[24928]: I1205 10:58:39.338312 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="e285e7ab-e929-426a-9d59-90a1382339e2" containerName="extract" Dec 05 10:58:39.344863 master-0 kubenswrapper[24928]: I1205 10:58:39.341870 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:39.346412 master-0 kubenswrapper[24928]: I1205 10:58:39.346365 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-2n9x6" Dec 05 10:58:39.361871 master-0 kubenswrapper[24928]: I1205 10:58:39.361675 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8"] Dec 05 10:58:39.437776 master-0 kubenswrapper[24928]: I1205 10:58:39.437704 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skspx\" (UniqueName: \"kubernetes.io/projected/d4719e26-d3cc-4ea3-804f-6d91530ea114-kube-api-access-skspx\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:39.438372 master-0 kubenswrapper[24928]: I1205 10:58:39.437844 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:39.438372 master-0 kubenswrapper[24928]: I1205 10:58:39.437892 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:39.539269 master-0 kubenswrapper[24928]: I1205 10:58:39.539176 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:39.539666 master-0 kubenswrapper[24928]: I1205 10:58:39.539454 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:39.539666 master-0 kubenswrapper[24928]: I1205 10:58:39.539532 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skspx\" (UniqueName: \"kubernetes.io/projected/d4719e26-d3cc-4ea3-804f-6d91530ea114-kube-api-access-skspx\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:39.539950 master-0 kubenswrapper[24928]: I1205 10:58:39.539903 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:39.540028 master-0 kubenswrapper[24928]: I1205 10:58:39.539975 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:39.561715 master-0 kubenswrapper[24928]: I1205 10:58:39.561634 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skspx\" (UniqueName: \"kubernetes.io/projected/d4719e26-d3cc-4ea3-804f-6d91530ea114-kube-api-access-skspx\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:39.671586 master-0 kubenswrapper[24928]: I1205 10:58:39.671443 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:40.076683 master-0 kubenswrapper[24928]: I1205 10:58:40.076616 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8"] Dec 05 10:58:40.079633 master-0 kubenswrapper[24928]: W1205 10:58:40.079583 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4719e26_d3cc_4ea3_804f_6d91530ea114.slice/crio-69f17689467a56bec67a55306b3b6f0a97ff156cc61257d1cb11f970f79383fa WatchSource:0}: Error finding container 69f17689467a56bec67a55306b3b6f0a97ff156cc61257d1cb11f970f79383fa: Status 404 returned error can't find the container with id 69f17689467a56bec67a55306b3b6f0a97ff156cc61257d1cb11f970f79383fa Dec 05 10:58:40.959452 master-0 kubenswrapper[24928]: I1205 10:58:40.958446 24928 generic.go:334] "Generic (PLEG): container finished" podID="d4719e26-d3cc-4ea3-804f-6d91530ea114" containerID="5ef4cc9e64676f133b76f8b3a96cfe39623a25f8790b7b8c49673c8782f5064f" exitCode=0 Dec 05 10:58:40.959452 master-0 kubenswrapper[24928]: I1205 10:58:40.958502 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" event={"ID":"d4719e26-d3cc-4ea3-804f-6d91530ea114","Type":"ContainerDied","Data":"5ef4cc9e64676f133b76f8b3a96cfe39623a25f8790b7b8c49673c8782f5064f"} Dec 05 10:58:40.959452 master-0 kubenswrapper[24928]: I1205 10:58:40.958538 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" event={"ID":"d4719e26-d3cc-4ea3-804f-6d91530ea114","Type":"ContainerStarted","Data":"69f17689467a56bec67a55306b3b6f0a97ff156cc61257d1cb11f970f79383fa"} Dec 05 10:58:41.498693 master-0 kubenswrapper[24928]: I1205 10:58:41.498612 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29"] Dec 05 10:58:41.499544 master-0 kubenswrapper[24928]: I1205 10:58:41.499512 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29" Dec 05 10:58:41.501278 master-0 kubenswrapper[24928]: I1205 10:58:41.501246 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 05 10:58:41.501858 master-0 kubenswrapper[24928]: I1205 10:58:41.501810 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 05 10:58:41.513223 master-0 kubenswrapper[24928]: I1205 10:58:41.513156 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29"] Dec 05 10:58:41.569344 master-0 kubenswrapper[24928]: I1205 10:58:41.569272 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnwgd\" (UniqueName: \"kubernetes.io/projected/198402ae-334b-4d09-8cd6-8218ff6cb79f-kube-api-access-bnwgd\") pod \"cert-manager-operator-controller-manager-64cf6dff88-5dc29\" (UID: \"198402ae-334b-4d09-8cd6-8218ff6cb79f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29" Dec 05 10:58:41.569344 master-0 kubenswrapper[24928]: I1205 10:58:41.569334 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/198402ae-334b-4d09-8cd6-8218ff6cb79f-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-5dc29\" (UID: \"198402ae-334b-4d09-8cd6-8218ff6cb79f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29" Dec 05 10:58:41.671296 master-0 kubenswrapper[24928]: I1205 10:58:41.671215 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnwgd\" (UniqueName: \"kubernetes.io/projected/198402ae-334b-4d09-8cd6-8218ff6cb79f-kube-api-access-bnwgd\") pod \"cert-manager-operator-controller-manager-64cf6dff88-5dc29\" (UID: \"198402ae-334b-4d09-8cd6-8218ff6cb79f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29" Dec 05 10:58:41.671296 master-0 kubenswrapper[24928]: I1205 10:58:41.671290 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/198402ae-334b-4d09-8cd6-8218ff6cb79f-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-5dc29\" (UID: \"198402ae-334b-4d09-8cd6-8218ff6cb79f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29" Dec 05 10:58:41.671886 master-0 kubenswrapper[24928]: I1205 10:58:41.671857 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/198402ae-334b-4d09-8cd6-8218ff6cb79f-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-5dc29\" (UID: \"198402ae-334b-4d09-8cd6-8218ff6cb79f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29" Dec 05 10:58:41.687579 master-0 kubenswrapper[24928]: I1205 10:58:41.687503 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnwgd\" (UniqueName: \"kubernetes.io/projected/198402ae-334b-4d09-8cd6-8218ff6cb79f-kube-api-access-bnwgd\") pod \"cert-manager-operator-controller-manager-64cf6dff88-5dc29\" (UID: \"198402ae-334b-4d09-8cd6-8218ff6cb79f\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29" Dec 05 10:58:41.832001 master-0 kubenswrapper[24928]: I1205 10:58:41.831921 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29" Dec 05 10:58:42.252718 master-0 kubenswrapper[24928]: I1205 10:58:42.252075 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29"] Dec 05 10:58:42.974859 master-0 kubenswrapper[24928]: I1205 10:58:42.974806 24928 generic.go:334] "Generic (PLEG): container finished" podID="d4719e26-d3cc-4ea3-804f-6d91530ea114" containerID="0e06bf7a6bcf6acb06cca2e33d75d3fefaf7be8c059eb4c7b911c75258583942" exitCode=0 Dec 05 10:58:42.975096 master-0 kubenswrapper[24928]: I1205 10:58:42.974849 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" event={"ID":"d4719e26-d3cc-4ea3-804f-6d91530ea114","Type":"ContainerDied","Data":"0e06bf7a6bcf6acb06cca2e33d75d3fefaf7be8c059eb4c7b911c75258583942"} Dec 05 10:58:42.976893 master-0 kubenswrapper[24928]: I1205 10:58:42.976747 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29" event={"ID":"198402ae-334b-4d09-8cd6-8218ff6cb79f","Type":"ContainerStarted","Data":"15c8ec9b427e3e93ec10b007dc4ebfc78d9517978eed11bc93ae4fd0c4c5f481"} Dec 05 10:58:43.990377 master-0 kubenswrapper[24928]: I1205 10:58:43.990308 24928 generic.go:334] "Generic (PLEG): container finished" podID="d4719e26-d3cc-4ea3-804f-6d91530ea114" containerID="ea8720e8a47701135b10a59cdfb91d9fa2c64940c76f6a5de8bbd0b98a91c47c" exitCode=0 Dec 05 10:58:43.990952 master-0 kubenswrapper[24928]: I1205 10:58:43.990371 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" event={"ID":"d4719e26-d3cc-4ea3-804f-6d91530ea114","Type":"ContainerDied","Data":"ea8720e8a47701135b10a59cdfb91d9fa2c64940c76f6a5de8bbd0b98a91c47c"} Dec 05 10:58:49.310313 master-0 kubenswrapper[24928]: I1205 10:58:49.310247 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:49.430147 master-0 kubenswrapper[24928]: I1205 10:58:49.429439 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-bundle\") pod \"d4719e26-d3cc-4ea3-804f-6d91530ea114\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " Dec 05 10:58:49.430147 master-0 kubenswrapper[24928]: I1205 10:58:49.429525 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skspx\" (UniqueName: \"kubernetes.io/projected/d4719e26-d3cc-4ea3-804f-6d91530ea114-kube-api-access-skspx\") pod \"d4719e26-d3cc-4ea3-804f-6d91530ea114\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " Dec 05 10:58:49.430147 master-0 kubenswrapper[24928]: I1205 10:58:49.429641 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-util\") pod \"d4719e26-d3cc-4ea3-804f-6d91530ea114\" (UID: \"d4719e26-d3cc-4ea3-804f-6d91530ea114\") " Dec 05 10:58:49.431478 master-0 kubenswrapper[24928]: I1205 10:58:49.431438 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-bundle" (OuterVolumeSpecName: "bundle") pod "d4719e26-d3cc-4ea3-804f-6d91530ea114" (UID: "d4719e26-d3cc-4ea3-804f-6d91530ea114"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:58:49.432609 master-0 kubenswrapper[24928]: I1205 10:58:49.432537 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4719e26-d3cc-4ea3-804f-6d91530ea114-kube-api-access-skspx" (OuterVolumeSpecName: "kube-api-access-skspx") pod "d4719e26-d3cc-4ea3-804f-6d91530ea114" (UID: "d4719e26-d3cc-4ea3-804f-6d91530ea114"). InnerVolumeSpecName "kube-api-access-skspx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:58:49.440503 master-0 kubenswrapper[24928]: I1205 10:58:49.440411 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-util" (OuterVolumeSpecName: "util") pod "d4719e26-d3cc-4ea3-804f-6d91530ea114" (UID: "d4719e26-d3cc-4ea3-804f-6d91530ea114"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:58:49.538791 master-0 kubenswrapper[24928]: I1205 10:58:49.533396 24928 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:49.538791 master-0 kubenswrapper[24928]: I1205 10:58:49.533480 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skspx\" (UniqueName: \"kubernetes.io/projected/d4719e26-d3cc-4ea3-804f-6d91530ea114-kube-api-access-skspx\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:49.538791 master-0 kubenswrapper[24928]: I1205 10:58:49.533496 24928 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d4719e26-d3cc-4ea3-804f-6d91530ea114-util\") on node \"master-0\" DevicePath \"\"" Dec 05 10:58:50.043406 master-0 kubenswrapper[24928]: I1205 10:58:50.043303 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" event={"ID":"d4719e26-d3cc-4ea3-804f-6d91530ea114","Type":"ContainerDied","Data":"69f17689467a56bec67a55306b3b6f0a97ff156cc61257d1cb11f970f79383fa"} Dec 05 10:58:50.043406 master-0 kubenswrapper[24928]: I1205 10:58:50.043363 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69f17689467a56bec67a55306b3b6f0a97ff156cc61257d1cb11f970f79383fa" Dec 05 10:58:50.043406 master-0 kubenswrapper[24928]: I1205 10:58:50.043373 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c92106nth8" Dec 05 10:58:50.045970 master-0 kubenswrapper[24928]: I1205 10:58:50.045857 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29" event={"ID":"198402ae-334b-4d09-8cd6-8218ff6cb79f","Type":"ContainerStarted","Data":"e32961aef2dbd22cafa65e23a22d1f47be1e514a0efcf3cd07c9fe23fc208dcc"} Dec 05 10:58:50.085309 master-0 kubenswrapper[24928]: I1205 10:58:50.085208 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-5dc29" podStartSLOduration=2.023326671 podStartE2EDuration="9.085190618s" podCreationTimestamp="2025-12-05 10:58:41 +0000 UTC" firstStartedPulling="2025-12-05 10:58:42.334086948 +0000 UTC m=+682.337280799" lastFinishedPulling="2025-12-05 10:58:49.395950895 +0000 UTC m=+689.399144746" observedRunningTime="2025-12-05 10:58:50.079255903 +0000 UTC m=+690.082449754" watchObservedRunningTime="2025-12-05 10:58:50.085190618 +0000 UTC m=+690.088384479" Dec 05 10:58:52.887589 master-0 kubenswrapper[24928]: I1205 10:58:52.887511 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-29nx4"] Dec 05 10:58:52.888265 master-0 kubenswrapper[24928]: E1205 10:58:52.887875 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4719e26-d3cc-4ea3-804f-6d91530ea114" containerName="util" Dec 05 10:58:52.888265 master-0 kubenswrapper[24928]: I1205 10:58:52.887895 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4719e26-d3cc-4ea3-804f-6d91530ea114" containerName="util" Dec 05 10:58:52.888265 master-0 kubenswrapper[24928]: E1205 10:58:52.887913 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4719e26-d3cc-4ea3-804f-6d91530ea114" containerName="extract" Dec 05 10:58:52.888265 master-0 kubenswrapper[24928]: I1205 10:58:52.887920 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4719e26-d3cc-4ea3-804f-6d91530ea114" containerName="extract" Dec 05 10:58:52.888265 master-0 kubenswrapper[24928]: E1205 10:58:52.887933 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4719e26-d3cc-4ea3-804f-6d91530ea114" containerName="pull" Dec 05 10:58:52.888265 master-0 kubenswrapper[24928]: I1205 10:58:52.887942 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4719e26-d3cc-4ea3-804f-6d91530ea114" containerName="pull" Dec 05 10:58:52.888265 master-0 kubenswrapper[24928]: I1205 10:58:52.888092 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4719e26-d3cc-4ea3-804f-6d91530ea114" containerName="extract" Dec 05 10:58:52.888752 master-0 kubenswrapper[24928]: I1205 10:58:52.888685 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" Dec 05 10:58:52.890942 master-0 kubenswrapper[24928]: I1205 10:58:52.890896 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 05 10:58:52.891088 master-0 kubenswrapper[24928]: I1205 10:58:52.891040 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 05 10:58:52.901149 master-0 kubenswrapper[24928]: I1205 10:58:52.901098 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-29nx4"] Dec 05 10:58:52.985872 master-0 kubenswrapper[24928]: I1205 10:58:52.985813 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc4ds\" (UniqueName: \"kubernetes.io/projected/5c4c235c-46d5-43e6-a12b-e93dd86a2fcf-kube-api-access-rc4ds\") pod \"cert-manager-webhook-f4fb5df64-29nx4\" (UID: \"5c4c235c-46d5-43e6-a12b-e93dd86a2fcf\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" Dec 05 10:58:52.986260 master-0 kubenswrapper[24928]: I1205 10:58:52.986240 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5c4c235c-46d5-43e6-a12b-e93dd86a2fcf-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-29nx4\" (UID: \"5c4c235c-46d5-43e6-a12b-e93dd86a2fcf\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" Dec 05 10:58:53.087355 master-0 kubenswrapper[24928]: I1205 10:58:53.087303 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5c4c235c-46d5-43e6-a12b-e93dd86a2fcf-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-29nx4\" (UID: \"5c4c235c-46d5-43e6-a12b-e93dd86a2fcf\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" Dec 05 10:58:53.087604 master-0 kubenswrapper[24928]: I1205 10:58:53.087394 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc4ds\" (UniqueName: \"kubernetes.io/projected/5c4c235c-46d5-43e6-a12b-e93dd86a2fcf-kube-api-access-rc4ds\") pod \"cert-manager-webhook-f4fb5df64-29nx4\" (UID: \"5c4c235c-46d5-43e6-a12b-e93dd86a2fcf\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" Dec 05 10:58:53.107658 master-0 kubenswrapper[24928]: I1205 10:58:53.107604 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc4ds\" (UniqueName: \"kubernetes.io/projected/5c4c235c-46d5-43e6-a12b-e93dd86a2fcf-kube-api-access-rc4ds\") pod \"cert-manager-webhook-f4fb5df64-29nx4\" (UID: \"5c4c235c-46d5-43e6-a12b-e93dd86a2fcf\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" Dec 05 10:58:53.113691 master-0 kubenswrapper[24928]: I1205 10:58:53.113648 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5c4c235c-46d5-43e6-a12b-e93dd86a2fcf-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-29nx4\" (UID: \"5c4c235c-46d5-43e6-a12b-e93dd86a2fcf\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" Dec 05 10:58:53.204244 master-0 kubenswrapper[24928]: I1205 10:58:53.204114 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" Dec 05 10:58:53.646637 master-0 kubenswrapper[24928]: I1205 10:58:53.646574 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-29nx4"] Dec 05 10:58:53.660561 master-0 kubenswrapper[24928]: W1205 10:58:53.660497 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c4c235c_46d5_43e6_a12b_e93dd86a2fcf.slice/crio-939ba3be77f42c9fb65fa1929eff2d5640cae4f97d5eb457334016edbd0a3216 WatchSource:0}: Error finding container 939ba3be77f42c9fb65fa1929eff2d5640cae4f97d5eb457334016edbd0a3216: Status 404 returned error can't find the container with id 939ba3be77f42c9fb65fa1929eff2d5640cae4f97d5eb457334016edbd0a3216 Dec 05 10:58:54.071634 master-0 kubenswrapper[24928]: I1205 10:58:54.071574 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" event={"ID":"5c4c235c-46d5-43e6-a12b-e93dd86a2fcf","Type":"ContainerStarted","Data":"939ba3be77f42c9fb65fa1929eff2d5640cae4f97d5eb457334016edbd0a3216"} Dec 05 10:58:55.218902 master-0 kubenswrapper[24928]: I1205 10:58:55.218823 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5fcz7"] Dec 05 10:58:55.220050 master-0 kubenswrapper[24928]: I1205 10:58:55.220013 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5fcz7" Dec 05 10:58:55.242647 master-0 kubenswrapper[24928]: I1205 10:58:55.241894 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 05 10:58:55.242647 master-0 kubenswrapper[24928]: I1205 10:58:55.242187 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 05 10:58:55.242647 master-0 kubenswrapper[24928]: I1205 10:58:55.242358 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5fcz7"] Dec 05 10:58:55.346999 master-0 kubenswrapper[24928]: I1205 10:58:55.346953 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2klk\" (UniqueName: \"kubernetes.io/projected/c79e22f5-57e7-437b-bdb4-52cc79cd73b0-kube-api-access-j2klk\") pod \"nmstate-operator-5b5b58f5c8-5fcz7\" (UID: \"c79e22f5-57e7-437b-bdb4-52cc79cd73b0\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5fcz7" Dec 05 10:58:55.450435 master-0 kubenswrapper[24928]: I1205 10:58:55.450323 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2klk\" (UniqueName: \"kubernetes.io/projected/c79e22f5-57e7-437b-bdb4-52cc79cd73b0-kube-api-access-j2klk\") pod \"nmstate-operator-5b5b58f5c8-5fcz7\" (UID: \"c79e22f5-57e7-437b-bdb4-52cc79cd73b0\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5fcz7" Dec 05 10:58:55.465907 master-0 kubenswrapper[24928]: I1205 10:58:55.465873 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2klk\" (UniqueName: \"kubernetes.io/projected/c79e22f5-57e7-437b-bdb4-52cc79cd73b0-kube-api-access-j2klk\") pod \"nmstate-operator-5b5b58f5c8-5fcz7\" (UID: \"c79e22f5-57e7-437b-bdb4-52cc79cd73b0\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5fcz7" Dec 05 10:58:55.559735 master-0 kubenswrapper[24928]: I1205 10:58:55.559544 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5fcz7" Dec 05 10:58:56.017413 master-0 kubenswrapper[24928]: I1205 10:58:56.017364 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5fcz7"] Dec 05 10:58:56.089363 master-0 kubenswrapper[24928]: I1205 10:58:56.089277 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5fcz7" event={"ID":"c79e22f5-57e7-437b-bdb4-52cc79cd73b0","Type":"ContainerStarted","Data":"af3d86dc04f4055cfceecf7170bfd214349f9854c747a1d5e377dace4c76cb11"} Dec 05 10:58:56.547798 master-0 kubenswrapper[24928]: I1205 10:58:56.544767 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-jkch2"] Dec 05 10:58:56.549607 master-0 kubenswrapper[24928]: I1205 10:58:56.548014 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jkch2" Dec 05 10:58:56.565177 master-0 kubenswrapper[24928]: I1205 10:58:56.565118 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-jkch2"] Dec 05 10:58:56.575162 master-0 kubenswrapper[24928]: I1205 10:58:56.575074 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dc0f03cc-abed-444e-8039-d6a9ed2c936b-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-jkch2\" (UID: \"dc0f03cc-abed-444e-8039-d6a9ed2c936b\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jkch2" Dec 05 10:58:56.575433 master-0 kubenswrapper[24928]: I1205 10:58:56.575381 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmr8d\" (UniqueName: \"kubernetes.io/projected/dc0f03cc-abed-444e-8039-d6a9ed2c936b-kube-api-access-wmr8d\") pod \"cert-manager-cainjector-855d9ccff4-jkch2\" (UID: \"dc0f03cc-abed-444e-8039-d6a9ed2c936b\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jkch2" Dec 05 10:58:56.676841 master-0 kubenswrapper[24928]: I1205 10:58:56.676770 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmr8d\" (UniqueName: \"kubernetes.io/projected/dc0f03cc-abed-444e-8039-d6a9ed2c936b-kube-api-access-wmr8d\") pod \"cert-manager-cainjector-855d9ccff4-jkch2\" (UID: \"dc0f03cc-abed-444e-8039-d6a9ed2c936b\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jkch2" Dec 05 10:58:56.677070 master-0 kubenswrapper[24928]: I1205 10:58:56.676852 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dc0f03cc-abed-444e-8039-d6a9ed2c936b-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-jkch2\" (UID: \"dc0f03cc-abed-444e-8039-d6a9ed2c936b\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jkch2" Dec 05 10:58:56.693015 master-0 kubenswrapper[24928]: I1205 10:58:56.692950 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dc0f03cc-abed-444e-8039-d6a9ed2c936b-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-jkch2\" (UID: \"dc0f03cc-abed-444e-8039-d6a9ed2c936b\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jkch2" Dec 05 10:58:56.696636 master-0 kubenswrapper[24928]: I1205 10:58:56.696599 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmr8d\" (UniqueName: \"kubernetes.io/projected/dc0f03cc-abed-444e-8039-d6a9ed2c936b-kube-api-access-wmr8d\") pod \"cert-manager-cainjector-855d9ccff4-jkch2\" (UID: \"dc0f03cc-abed-444e-8039-d6a9ed2c936b\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-jkch2" Dec 05 10:58:56.884231 master-0 kubenswrapper[24928]: I1205 10:58:56.884082 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jkch2" Dec 05 10:59:02.419464 master-0 kubenswrapper[24928]: I1205 10:59:02.415957 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-jkch2"] Dec 05 10:59:03.163970 master-0 kubenswrapper[24928]: I1205 10:59:03.163906 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jkch2" event={"ID":"dc0f03cc-abed-444e-8039-d6a9ed2c936b","Type":"ContainerStarted","Data":"0246b04594e5190cebea7101aa6736231e76572fa55dccd3c31c731646003819"} Dec 05 10:59:03.164243 master-0 kubenswrapper[24928]: I1205 10:59:03.163989 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jkch2" event={"ID":"dc0f03cc-abed-444e-8039-d6a9ed2c936b","Type":"ContainerStarted","Data":"743e076848c736a9d842bc58c648bbc19430c7b298e315dc861f73dec6d1688e"} Dec 05 10:59:03.168635 master-0 kubenswrapper[24928]: I1205 10:59:03.168531 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" event={"ID":"5c4c235c-46d5-43e6-a12b-e93dd86a2fcf","Type":"ContainerStarted","Data":"e6fa7065143297b6aef14bc963080718c4ce62bd785acdd3ae405b57bdcc66c2"} Dec 05 10:59:03.168766 master-0 kubenswrapper[24928]: I1205 10:59:03.168677 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" Dec 05 10:59:03.186770 master-0 kubenswrapper[24928]: I1205 10:59:03.186700 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-jkch2" podStartSLOduration=7.18668104 podStartE2EDuration="7.18668104s" podCreationTimestamp="2025-12-05 10:58:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:59:03.182138178 +0000 UTC m=+703.185332029" watchObservedRunningTime="2025-12-05 10:59:03.18668104 +0000 UTC m=+703.189874891" Dec 05 10:59:03.202590 master-0 kubenswrapper[24928]: I1205 10:59:03.202280 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" podStartSLOduration=2.8194840279999998 podStartE2EDuration="11.202261001s" podCreationTimestamp="2025-12-05 10:58:52 +0000 UTC" firstStartedPulling="2025-12-05 10:58:53.663950512 +0000 UTC m=+693.667144363" lastFinishedPulling="2025-12-05 10:59:02.046727485 +0000 UTC m=+702.049921336" observedRunningTime="2025-12-05 10:59:03.200947559 +0000 UTC m=+703.204141420" watchObservedRunningTime="2025-12-05 10:59:03.202261001 +0000 UTC m=+703.205454862" Dec 05 10:59:04.118363 master-0 kubenswrapper[24928]: I1205 10:59:04.118298 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m"] Dec 05 10:59:04.119259 master-0 kubenswrapper[24928]: I1205 10:59:04.119232 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:04.130628 master-0 kubenswrapper[24928]: I1205 10:59:04.130540 24928 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 05 10:59:04.130850 master-0 kubenswrapper[24928]: I1205 10:59:04.130706 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 05 10:59:04.130850 master-0 kubenswrapper[24928]: I1205 10:59:04.130786 24928 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 05 10:59:04.130975 master-0 kubenswrapper[24928]: I1205 10:59:04.130907 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 05 10:59:04.178833 master-0 kubenswrapper[24928]: I1205 10:59:04.178766 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5fcz7" event={"ID":"c79e22f5-57e7-437b-bdb4-52cc79cd73b0","Type":"ContainerStarted","Data":"e5b33d4fcbee5e27461382a6b4263340afb695dddc9dfb328a76d6a4409b54e3"} Dec 05 10:59:04.224794 master-0 kubenswrapper[24928]: I1205 10:59:04.224688 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m"] Dec 05 10:59:04.277447 master-0 kubenswrapper[24928]: I1205 10:59:04.269399 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmrqw\" (UniqueName: \"kubernetes.io/projected/697f986f-bfa7-4e85-9d4b-5f07a4675228-kube-api-access-fmrqw\") pod \"metallb-operator-controller-manager-9d5bd9bc7-q878m\" (UID: \"697f986f-bfa7-4e85-9d4b-5f07a4675228\") " pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:04.277447 master-0 kubenswrapper[24928]: I1205 10:59:04.269532 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/697f986f-bfa7-4e85-9d4b-5f07a4675228-webhook-cert\") pod \"metallb-operator-controller-manager-9d5bd9bc7-q878m\" (UID: \"697f986f-bfa7-4e85-9d4b-5f07a4675228\") " pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:04.277447 master-0 kubenswrapper[24928]: I1205 10:59:04.269556 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/697f986f-bfa7-4e85-9d4b-5f07a4675228-apiservice-cert\") pod \"metallb-operator-controller-manager-9d5bd9bc7-q878m\" (UID: \"697f986f-bfa7-4e85-9d4b-5f07a4675228\") " pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:04.372504 master-0 kubenswrapper[24928]: I1205 10:59:04.371967 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmrqw\" (UniqueName: \"kubernetes.io/projected/697f986f-bfa7-4e85-9d4b-5f07a4675228-kube-api-access-fmrqw\") pod \"metallb-operator-controller-manager-9d5bd9bc7-q878m\" (UID: \"697f986f-bfa7-4e85-9d4b-5f07a4675228\") " pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:04.372504 master-0 kubenswrapper[24928]: I1205 10:59:04.372066 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/697f986f-bfa7-4e85-9d4b-5f07a4675228-webhook-cert\") pod \"metallb-operator-controller-manager-9d5bd9bc7-q878m\" (UID: \"697f986f-bfa7-4e85-9d4b-5f07a4675228\") " pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:04.372504 master-0 kubenswrapper[24928]: I1205 10:59:04.372086 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/697f986f-bfa7-4e85-9d4b-5f07a4675228-apiservice-cert\") pod \"metallb-operator-controller-manager-9d5bd9bc7-q878m\" (UID: \"697f986f-bfa7-4e85-9d4b-5f07a4675228\") " pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:04.385444 master-0 kubenswrapper[24928]: I1205 10:59:04.382411 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/697f986f-bfa7-4e85-9d4b-5f07a4675228-webhook-cert\") pod \"metallb-operator-controller-manager-9d5bd9bc7-q878m\" (UID: \"697f986f-bfa7-4e85-9d4b-5f07a4675228\") " pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:04.385444 master-0 kubenswrapper[24928]: I1205 10:59:04.383225 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/697f986f-bfa7-4e85-9d4b-5f07a4675228-apiservice-cert\") pod \"metallb-operator-controller-manager-9d5bd9bc7-q878m\" (UID: \"697f986f-bfa7-4e85-9d4b-5f07a4675228\") " pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:04.385444 master-0 kubenswrapper[24928]: I1205 10:59:04.384306 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5fcz7" podStartSLOduration=1.944816639 podStartE2EDuration="9.384295536s" podCreationTimestamp="2025-12-05 10:58:55 +0000 UTC" firstStartedPulling="2025-12-05 10:58:56.014556012 +0000 UTC m=+696.017749863" lastFinishedPulling="2025-12-05 10:59:03.454034909 +0000 UTC m=+703.457228760" observedRunningTime="2025-12-05 10:59:04.372123548 +0000 UTC m=+704.375317409" watchObservedRunningTime="2025-12-05 10:59:04.384295536 +0000 UTC m=+704.387489377" Dec 05 10:59:04.424524 master-0 kubenswrapper[24928]: I1205 10:59:04.424443 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmrqw\" (UniqueName: \"kubernetes.io/projected/697f986f-bfa7-4e85-9d4b-5f07a4675228-kube-api-access-fmrqw\") pod \"metallb-operator-controller-manager-9d5bd9bc7-q878m\" (UID: \"697f986f-bfa7-4e85-9d4b-5f07a4675228\") " pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:04.437746 master-0 kubenswrapper[24928]: I1205 10:59:04.437660 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:04.705670 master-0 kubenswrapper[24928]: I1205 10:59:04.704316 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x"] Dec 05 10:59:04.705670 master-0 kubenswrapper[24928]: I1205 10:59:04.705320 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:04.708630 master-0 kubenswrapper[24928]: I1205 10:59:04.708575 24928 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 10:59:04.708949 master-0 kubenswrapper[24928]: I1205 10:59:04.708915 24928 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 05 10:59:04.730443 master-0 kubenswrapper[24928]: I1205 10:59:04.725480 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x"] Dec 05 10:59:04.904452 master-0 kubenswrapper[24928]: I1205 10:59:04.898446 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c39fe00d-1550-4d88-8b8a-95439d6e58e1-apiservice-cert\") pod \"metallb-operator-webhook-server-5f77dd7bb4-xmg4x\" (UID: \"c39fe00d-1550-4d88-8b8a-95439d6e58e1\") " pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:04.904452 master-0 kubenswrapper[24928]: I1205 10:59:04.898567 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c39fe00d-1550-4d88-8b8a-95439d6e58e1-webhook-cert\") pod \"metallb-operator-webhook-server-5f77dd7bb4-xmg4x\" (UID: \"c39fe00d-1550-4d88-8b8a-95439d6e58e1\") " pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:04.904452 master-0 kubenswrapper[24928]: I1205 10:59:04.898634 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4qk6\" (UniqueName: \"kubernetes.io/projected/c39fe00d-1550-4d88-8b8a-95439d6e58e1-kube-api-access-b4qk6\") pod \"metallb-operator-webhook-server-5f77dd7bb4-xmg4x\" (UID: \"c39fe00d-1550-4d88-8b8a-95439d6e58e1\") " pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:04.975822 master-0 kubenswrapper[24928]: I1205 10:59:04.974112 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m"] Dec 05 10:59:05.004481 master-0 kubenswrapper[24928]: I1205 10:59:05.001343 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c39fe00d-1550-4d88-8b8a-95439d6e58e1-webhook-cert\") pod \"metallb-operator-webhook-server-5f77dd7bb4-xmg4x\" (UID: \"c39fe00d-1550-4d88-8b8a-95439d6e58e1\") " pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:05.004481 master-0 kubenswrapper[24928]: I1205 10:59:05.001502 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4qk6\" (UniqueName: \"kubernetes.io/projected/c39fe00d-1550-4d88-8b8a-95439d6e58e1-kube-api-access-b4qk6\") pod \"metallb-operator-webhook-server-5f77dd7bb4-xmg4x\" (UID: \"c39fe00d-1550-4d88-8b8a-95439d6e58e1\") " pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:05.004481 master-0 kubenswrapper[24928]: I1205 10:59:05.001551 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c39fe00d-1550-4d88-8b8a-95439d6e58e1-apiservice-cert\") pod \"metallb-operator-webhook-server-5f77dd7bb4-xmg4x\" (UID: \"c39fe00d-1550-4d88-8b8a-95439d6e58e1\") " pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:05.012494 master-0 kubenswrapper[24928]: I1205 10:59:05.010341 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c39fe00d-1550-4d88-8b8a-95439d6e58e1-apiservice-cert\") pod \"metallb-operator-webhook-server-5f77dd7bb4-xmg4x\" (UID: \"c39fe00d-1550-4d88-8b8a-95439d6e58e1\") " pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:05.016443 master-0 kubenswrapper[24928]: I1205 10:59:05.013337 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c39fe00d-1550-4d88-8b8a-95439d6e58e1-webhook-cert\") pod \"metallb-operator-webhook-server-5f77dd7bb4-xmg4x\" (UID: \"c39fe00d-1550-4d88-8b8a-95439d6e58e1\") " pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:05.037318 master-0 kubenswrapper[24928]: I1205 10:59:05.037266 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4qk6\" (UniqueName: \"kubernetes.io/projected/c39fe00d-1550-4d88-8b8a-95439d6e58e1-kube-api-access-b4qk6\") pod \"metallb-operator-webhook-server-5f77dd7bb4-xmg4x\" (UID: \"c39fe00d-1550-4d88-8b8a-95439d6e58e1\") " pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:05.076448 master-0 kubenswrapper[24928]: I1205 10:59:05.063920 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:05.198664 master-0 kubenswrapper[24928]: I1205 10:59:05.195582 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" event={"ID":"697f986f-bfa7-4e85-9d4b-5f07a4675228","Type":"ContainerStarted","Data":"c1cab48c1c952a5faf1b6948e650d1b501a3a962eaaabe5a7b417c2c1619e293"} Dec 05 10:59:05.634401 master-0 kubenswrapper[24928]: I1205 10:59:05.634315 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x"] Dec 05 10:59:06.213664 master-0 kubenswrapper[24928]: I1205 10:59:06.213586 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" event={"ID":"c39fe00d-1550-4d88-8b8a-95439d6e58e1","Type":"ContainerStarted","Data":"63d432a860e9e9287181028217630d0769cda5d32d82b8c81aa7484ae5474191"} Dec 05 10:59:08.224576 master-0 kubenswrapper[24928]: I1205 10:59:08.222360 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-29nx4" Dec 05 10:59:10.270792 master-0 kubenswrapper[24928]: I1205 10:59:10.268852 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" event={"ID":"697f986f-bfa7-4e85-9d4b-5f07a4675228","Type":"ContainerStarted","Data":"e9486124d13bc8dae987c24070ca54a94739e7fe1ad33c970e9df6849efa6c6c"} Dec 05 10:59:10.270792 master-0 kubenswrapper[24928]: I1205 10:59:10.268933 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:10.312339 master-0 kubenswrapper[24928]: I1205 10:59:10.312202 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" podStartSLOduration=2.127115448 podStartE2EDuration="6.312170824s" podCreationTimestamp="2025-12-05 10:59:04 +0000 UTC" firstStartedPulling="2025-12-05 10:59:04.990077026 +0000 UTC m=+704.993270877" lastFinishedPulling="2025-12-05 10:59:09.175132402 +0000 UTC m=+709.178326253" observedRunningTime="2025-12-05 10:59:10.296402957 +0000 UTC m=+710.299596818" watchObservedRunningTime="2025-12-05 10:59:10.312170824 +0000 UTC m=+710.315364665" Dec 05 10:59:11.135438 master-0 kubenswrapper[24928]: I1205 10:59:11.134963 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-nj5nk"] Dec 05 10:59:11.140414 master-0 kubenswrapper[24928]: I1205 10:59:11.136246 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-nj5nk" Dec 05 10:59:11.140414 master-0 kubenswrapper[24928]: I1205 10:59:11.139306 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 05 10:59:11.140414 master-0 kubenswrapper[24928]: I1205 10:59:11.139537 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 05 10:59:11.164892 master-0 kubenswrapper[24928]: I1205 10:59:11.161543 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-nj5nk"] Dec 05 10:59:11.168858 master-0 kubenswrapper[24928]: I1205 10:59:11.166122 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7t6h4\" (UniqueName: \"kubernetes.io/projected/0962dd5b-ecea-4032-84e5-d848b7e46451-kube-api-access-7t6h4\") pod \"obo-prometheus-operator-668cf9dfbb-nj5nk\" (UID: \"0962dd5b-ecea-4032-84e5-d848b7e46451\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-nj5nk" Dec 05 10:59:11.275458 master-0 kubenswrapper[24928]: I1205 10:59:11.271366 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7t6h4\" (UniqueName: \"kubernetes.io/projected/0962dd5b-ecea-4032-84e5-d848b7e46451-kube-api-access-7t6h4\") pod \"obo-prometheus-operator-668cf9dfbb-nj5nk\" (UID: \"0962dd5b-ecea-4032-84e5-d848b7e46451\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-nj5nk" Dec 05 10:59:12.509641 master-0 kubenswrapper[24928]: I1205 10:59:12.509512 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-4l8x5"] Dec 05 10:59:12.510663 master-0 kubenswrapper[24928]: I1205 10:59:12.510615 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-4l8x5" Dec 05 10:59:12.696678 master-0 kubenswrapper[24928]: I1205 10:59:12.696615 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x24qx\" (UniqueName: \"kubernetes.io/projected/f22dc7c0-3179-4346-8a71-865c2a0bb10e-kube-api-access-x24qx\") pod \"cert-manager-86cb77c54b-4l8x5\" (UID: \"f22dc7c0-3179-4346-8a71-865c2a0bb10e\") " pod="cert-manager/cert-manager-86cb77c54b-4l8x5" Dec 05 10:59:12.696958 master-0 kubenswrapper[24928]: I1205 10:59:12.696940 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f22dc7c0-3179-4346-8a71-865c2a0bb10e-bound-sa-token\") pod \"cert-manager-86cb77c54b-4l8x5\" (UID: \"f22dc7c0-3179-4346-8a71-865c2a0bb10e\") " pod="cert-manager/cert-manager-86cb77c54b-4l8x5" Dec 05 10:59:12.798769 master-0 kubenswrapper[24928]: I1205 10:59:12.798672 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x24qx\" (UniqueName: \"kubernetes.io/projected/f22dc7c0-3179-4346-8a71-865c2a0bb10e-kube-api-access-x24qx\") pod \"cert-manager-86cb77c54b-4l8x5\" (UID: \"f22dc7c0-3179-4346-8a71-865c2a0bb10e\") " pod="cert-manager/cert-manager-86cb77c54b-4l8x5" Dec 05 10:59:12.798769 master-0 kubenswrapper[24928]: I1205 10:59:12.798758 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f22dc7c0-3179-4346-8a71-865c2a0bb10e-bound-sa-token\") pod \"cert-manager-86cb77c54b-4l8x5\" (UID: \"f22dc7c0-3179-4346-8a71-865c2a0bb10e\") " pod="cert-manager/cert-manager-86cb77c54b-4l8x5" Dec 05 10:59:15.479451 master-0 kubenswrapper[24928]: I1205 10:59:15.478217 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29"] Dec 05 10:59:15.480225 master-0 kubenswrapper[24928]: I1205 10:59:15.479521 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29" Dec 05 10:59:15.484181 master-0 kubenswrapper[24928]: I1205 10:59:15.480408 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7t6h4\" (UniqueName: \"kubernetes.io/projected/0962dd5b-ecea-4032-84e5-d848b7e46451-kube-api-access-7t6h4\") pod \"obo-prometheus-operator-668cf9dfbb-nj5nk\" (UID: \"0962dd5b-ecea-4032-84e5-d848b7e46451\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-nj5nk" Dec 05 10:59:15.484181 master-0 kubenswrapper[24928]: I1205 10:59:15.482674 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 05 10:59:15.504721 master-0 kubenswrapper[24928]: I1205 10:59:15.504491 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x24qx\" (UniqueName: \"kubernetes.io/projected/f22dc7c0-3179-4346-8a71-865c2a0bb10e-kube-api-access-x24qx\") pod \"cert-manager-86cb77c54b-4l8x5\" (UID: \"f22dc7c0-3179-4346-8a71-865c2a0bb10e\") " pod="cert-manager/cert-manager-86cb77c54b-4l8x5" Dec 05 10:59:15.517585 master-0 kubenswrapper[24928]: I1205 10:59:15.512817 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f22dc7c0-3179-4346-8a71-865c2a0bb10e-bound-sa-token\") pod \"cert-manager-86cb77c54b-4l8x5\" (UID: \"f22dc7c0-3179-4346-8a71-865c2a0bb10e\") " pod="cert-manager/cert-manager-86cb77c54b-4l8x5" Dec 05 10:59:15.530946 master-0 kubenswrapper[24928]: I1205 10:59:15.529023 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-4l8x5" Dec 05 10:59:15.543458 master-0 kubenswrapper[24928]: I1205 10:59:15.541634 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-4l8x5"] Dec 05 10:59:15.572501 master-0 kubenswrapper[24928]: I1205 10:59:15.572045 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29"] Dec 05 10:59:15.583522 master-0 kubenswrapper[24928]: I1205 10:59:15.583457 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz"] Dec 05 10:59:15.585332 master-0 kubenswrapper[24928]: I1205 10:59:15.585275 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz" Dec 05 10:59:15.637624 master-0 kubenswrapper[24928]: I1205 10:59:15.637541 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz"] Dec 05 10:59:15.661631 master-0 kubenswrapper[24928]: I1205 10:59:15.661037 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-nj5nk" Dec 05 10:59:15.675900 master-0 kubenswrapper[24928]: I1205 10:59:15.674770 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/686c75b7-f169-41df-aa2e-80f48a404f30-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-zpw29\" (UID: \"686c75b7-f169-41df-aa2e-80f48a404f30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29" Dec 05 10:59:15.675900 master-0 kubenswrapper[24928]: I1205 10:59:15.674886 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/686c75b7-f169-41df-aa2e-80f48a404f30-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-zpw29\" (UID: \"686c75b7-f169-41df-aa2e-80f48a404f30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29" Dec 05 10:59:15.692619 master-0 kubenswrapper[24928]: I1205 10:59:15.683911 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-g242x"] Dec 05 10:59:15.692619 master-0 kubenswrapper[24928]: I1205 10:59:15.685001 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-g242x" Dec 05 10:59:15.692619 master-0 kubenswrapper[24928]: I1205 10:59:15.688213 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 05 10:59:15.723365 master-0 kubenswrapper[24928]: I1205 10:59:15.710035 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-g242x"] Dec 05 10:59:15.778460 master-0 kubenswrapper[24928]: I1205 10:59:15.776772 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c2e6173-1e2a-48ef-94ae-70974b457589-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-l52lz\" (UID: \"0c2e6173-1e2a-48ef-94ae-70974b457589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz" Dec 05 10:59:15.778460 master-0 kubenswrapper[24928]: I1205 10:59:15.776906 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/686c75b7-f169-41df-aa2e-80f48a404f30-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-zpw29\" (UID: \"686c75b7-f169-41df-aa2e-80f48a404f30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29" Dec 05 10:59:15.778460 master-0 kubenswrapper[24928]: I1205 10:59:15.776993 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/686c75b7-f169-41df-aa2e-80f48a404f30-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-zpw29\" (UID: \"686c75b7-f169-41df-aa2e-80f48a404f30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29" Dec 05 10:59:15.778460 master-0 kubenswrapper[24928]: I1205 10:59:15.777025 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c2e6173-1e2a-48ef-94ae-70974b457589-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-l52lz\" (UID: \"0c2e6173-1e2a-48ef-94ae-70974b457589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz" Dec 05 10:59:15.783810 master-0 kubenswrapper[24928]: I1205 10:59:15.782268 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/686c75b7-f169-41df-aa2e-80f48a404f30-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-zpw29\" (UID: \"686c75b7-f169-41df-aa2e-80f48a404f30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29" Dec 05 10:59:15.789986 master-0 kubenswrapper[24928]: I1205 10:59:15.789927 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/686c75b7-f169-41df-aa2e-80f48a404f30-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-zpw29\" (UID: \"686c75b7-f169-41df-aa2e-80f48a404f30\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29" Dec 05 10:59:15.880299 master-0 kubenswrapper[24928]: I1205 10:59:15.880220 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c2e6173-1e2a-48ef-94ae-70974b457589-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-l52lz\" (UID: \"0c2e6173-1e2a-48ef-94ae-70974b457589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz" Dec 05 10:59:15.880666 master-0 kubenswrapper[24928]: I1205 10:59:15.880366 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c2e6173-1e2a-48ef-94ae-70974b457589-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-l52lz\" (UID: \"0c2e6173-1e2a-48ef-94ae-70974b457589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz" Dec 05 10:59:15.880666 master-0 kubenswrapper[24928]: I1205 10:59:15.880414 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c56gm\" (UniqueName: \"kubernetes.io/projected/4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6-kube-api-access-c56gm\") pod \"observability-operator-d8bb48f5d-g242x\" (UID: \"4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6\") " pod="openshift-operators/observability-operator-d8bb48f5d-g242x" Dec 05 10:59:15.880666 master-0 kubenswrapper[24928]: I1205 10:59:15.880472 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-g242x\" (UID: \"4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6\") " pod="openshift-operators/observability-operator-d8bb48f5d-g242x" Dec 05 10:59:15.887347 master-0 kubenswrapper[24928]: I1205 10:59:15.887206 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0c2e6173-1e2a-48ef-94ae-70974b457589-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-l52lz\" (UID: \"0c2e6173-1e2a-48ef-94ae-70974b457589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz" Dec 05 10:59:15.893460 master-0 kubenswrapper[24928]: I1205 10:59:15.893366 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0c2e6173-1e2a-48ef-94ae-70974b457589-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-78b56678b9-l52lz\" (UID: \"0c2e6173-1e2a-48ef-94ae-70974b457589\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz" Dec 05 10:59:15.906394 master-0 kubenswrapper[24928]: I1205 10:59:15.906328 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29" Dec 05 10:59:15.926461 master-0 kubenswrapper[24928]: I1205 10:59:15.921672 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-jw8mn"] Dec 05 10:59:15.926461 master-0 kubenswrapper[24928]: I1205 10:59:15.923588 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-jw8mn" Dec 05 10:59:15.931548 master-0 kubenswrapper[24928]: I1205 10:59:15.931460 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-jw8mn"] Dec 05 10:59:16.001523 master-0 kubenswrapper[24928]: I1205 10:59:15.997934 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c56gm\" (UniqueName: \"kubernetes.io/projected/4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6-kube-api-access-c56gm\") pod \"observability-operator-d8bb48f5d-g242x\" (UID: \"4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6\") " pod="openshift-operators/observability-operator-d8bb48f5d-g242x" Dec 05 10:59:16.001523 master-0 kubenswrapper[24928]: I1205 10:59:15.998005 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-g242x\" (UID: \"4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6\") " pod="openshift-operators/observability-operator-d8bb48f5d-g242x" Dec 05 10:59:16.001523 master-0 kubenswrapper[24928]: I1205 10:59:15.998031 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zlxr\" (UniqueName: \"kubernetes.io/projected/7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6-kube-api-access-4zlxr\") pod \"perses-operator-5446b9c989-jw8mn\" (UID: \"7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6\") " pod="openshift-operators/perses-operator-5446b9c989-jw8mn" Dec 05 10:59:16.001523 master-0 kubenswrapper[24928]: I1205 10:59:15.998123 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6-openshift-service-ca\") pod \"perses-operator-5446b9c989-jw8mn\" (UID: \"7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6\") " pod="openshift-operators/perses-operator-5446b9c989-jw8mn" Dec 05 10:59:16.011553 master-0 kubenswrapper[24928]: I1205 10:59:16.009763 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-g242x\" (UID: \"4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6\") " pod="openshift-operators/observability-operator-d8bb48f5d-g242x" Dec 05 10:59:16.021811 master-0 kubenswrapper[24928]: I1205 10:59:16.021753 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c56gm\" (UniqueName: \"kubernetes.io/projected/4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6-kube-api-access-c56gm\") pod \"observability-operator-d8bb48f5d-g242x\" (UID: \"4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6\") " pod="openshift-operators/observability-operator-d8bb48f5d-g242x" Dec 05 10:59:16.028604 master-0 kubenswrapper[24928]: I1205 10:59:16.028543 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz" Dec 05 10:59:16.102504 master-0 kubenswrapper[24928]: I1205 10:59:16.102415 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6-openshift-service-ca\") pod \"perses-operator-5446b9c989-jw8mn\" (UID: \"7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6\") " pod="openshift-operators/perses-operator-5446b9c989-jw8mn" Dec 05 10:59:16.102807 master-0 kubenswrapper[24928]: I1205 10:59:16.102785 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zlxr\" (UniqueName: \"kubernetes.io/projected/7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6-kube-api-access-4zlxr\") pod \"perses-operator-5446b9c989-jw8mn\" (UID: \"7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6\") " pod="openshift-operators/perses-operator-5446b9c989-jw8mn" Dec 05 10:59:16.122453 master-0 kubenswrapper[24928]: I1205 10:59:16.104184 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6-openshift-service-ca\") pod \"perses-operator-5446b9c989-jw8mn\" (UID: \"7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6\") " pod="openshift-operators/perses-operator-5446b9c989-jw8mn" Dec 05 10:59:16.122828 master-0 kubenswrapper[24928]: I1205 10:59:16.117757 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-g242x" Dec 05 10:59:16.159447 master-0 kubenswrapper[24928]: I1205 10:59:16.157375 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zlxr\" (UniqueName: \"kubernetes.io/projected/7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6-kube-api-access-4zlxr\") pod \"perses-operator-5446b9c989-jw8mn\" (UID: \"7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6\") " pod="openshift-operators/perses-operator-5446b9c989-jw8mn" Dec 05 10:59:16.285702 master-0 kubenswrapper[24928]: I1205 10:59:16.285654 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-jw8mn" Dec 05 10:59:16.359684 master-0 kubenswrapper[24928]: I1205 10:59:16.357019 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-4l8x5"] Dec 05 10:59:16.359684 master-0 kubenswrapper[24928]: I1205 10:59:16.357078 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-nj5nk"] Dec 05 10:59:16.359684 master-0 kubenswrapper[24928]: I1205 10:59:16.359291 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" event={"ID":"c39fe00d-1550-4d88-8b8a-95439d6e58e1","Type":"ContainerStarted","Data":"fee0cf2ad6dce23badde610ba684141a8e26ec8e62db2ac6cd6b70d548d52f2f"} Dec 05 10:59:16.360200 master-0 kubenswrapper[24928]: I1205 10:59:16.360154 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:16.389834 master-0 kubenswrapper[24928]: W1205 10:59:16.388546 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf22dc7c0_3179_4346_8a71_865c2a0bb10e.slice/crio-14f355ae641d6fcc88df64339a82b3824e716c50ee2c806b780eadd8d660f009 WatchSource:0}: Error finding container 14f355ae641d6fcc88df64339a82b3824e716c50ee2c806b780eadd8d660f009: Status 404 returned error can't find the container with id 14f355ae641d6fcc88df64339a82b3824e716c50ee2c806b780eadd8d660f009 Dec 05 10:59:16.389834 master-0 kubenswrapper[24928]: W1205 10:59:16.389047 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0962dd5b_ecea_4032_84e5_d848b7e46451.slice/crio-a27e8f74b419774b99e570ca0e40bc5e247a880637d3aaa6945295147420cf98 WatchSource:0}: Error finding container a27e8f74b419774b99e570ca0e40bc5e247a880637d3aaa6945295147420cf98: Status 404 returned error can't find the container with id a27e8f74b419774b99e570ca0e40bc5e247a880637d3aaa6945295147420cf98 Dec 05 10:59:16.411002 master-0 kubenswrapper[24928]: I1205 10:59:16.410861 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" podStartSLOduration=2.410080418 podStartE2EDuration="12.410557489s" podCreationTimestamp="2025-12-05 10:59:04 +0000 UTC" firstStartedPulling="2025-12-05 10:59:05.647812406 +0000 UTC m=+705.651006257" lastFinishedPulling="2025-12-05 10:59:15.648289477 +0000 UTC m=+715.651483328" observedRunningTime="2025-12-05 10:59:16.385902796 +0000 UTC m=+716.389096657" watchObservedRunningTime="2025-12-05 10:59:16.410557489 +0000 UTC m=+716.413751330" Dec 05 10:59:16.719678 master-0 kubenswrapper[24928]: I1205 10:59:16.719620 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz"] Dec 05 10:59:16.805817 master-0 kubenswrapper[24928]: I1205 10:59:16.802142 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29"] Dec 05 10:59:16.814838 master-0 kubenswrapper[24928]: W1205 10:59:16.814737 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod686c75b7_f169_41df_aa2e_80f48a404f30.slice/crio-5244c2d9a86ad6a2b8e351661c61ab880e20b154c8091971afb0bc9aade659b2 WatchSource:0}: Error finding container 5244c2d9a86ad6a2b8e351661c61ab880e20b154c8091971afb0bc9aade659b2: Status 404 returned error can't find the container with id 5244c2d9a86ad6a2b8e351661c61ab880e20b154c8091971afb0bc9aade659b2 Dec 05 10:59:16.881179 master-0 kubenswrapper[24928]: I1205 10:59:16.881106 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-g242x"] Dec 05 10:59:16.881375 master-0 kubenswrapper[24928]: W1205 10:59:16.881214 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f87e93b_6e0f_4ecf_80f7_46c3fd613ae6.slice/crio-4860e5e440bbbcffda63a054bf6ce9af0e311c31872f2cd3503d3b04bc360789 WatchSource:0}: Error finding container 4860e5e440bbbcffda63a054bf6ce9af0e311c31872f2cd3503d3b04bc360789: Status 404 returned error can't find the container with id 4860e5e440bbbcffda63a054bf6ce9af0e311c31872f2cd3503d3b04bc360789 Dec 05 10:59:16.968473 master-0 kubenswrapper[24928]: I1205 10:59:16.968413 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-jw8mn"] Dec 05 10:59:17.377833 master-0 kubenswrapper[24928]: I1205 10:59:17.377711 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-g242x" event={"ID":"4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6","Type":"ContainerStarted","Data":"4860e5e440bbbcffda63a054bf6ce9af0e311c31872f2cd3503d3b04bc360789"} Dec 05 10:59:17.379779 master-0 kubenswrapper[24928]: I1205 10:59:17.379736 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-4l8x5" event={"ID":"f22dc7c0-3179-4346-8a71-865c2a0bb10e","Type":"ContainerStarted","Data":"be3fc75ff42bc7afb4174363d4e6988439a34b6ad8e2ea201b88210083d0a094"} Dec 05 10:59:17.379779 master-0 kubenswrapper[24928]: I1205 10:59:17.379771 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-4l8x5" event={"ID":"f22dc7c0-3179-4346-8a71-865c2a0bb10e","Type":"ContainerStarted","Data":"14f355ae641d6fcc88df64339a82b3824e716c50ee2c806b780eadd8d660f009"} Dec 05 10:59:17.381251 master-0 kubenswrapper[24928]: I1205 10:59:17.381202 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29" event={"ID":"686c75b7-f169-41df-aa2e-80f48a404f30","Type":"ContainerStarted","Data":"5244c2d9a86ad6a2b8e351661c61ab880e20b154c8091971afb0bc9aade659b2"} Dec 05 10:59:17.382826 master-0 kubenswrapper[24928]: I1205 10:59:17.382713 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-jw8mn" event={"ID":"7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6","Type":"ContainerStarted","Data":"6888a1892d910c6cac72ca5c4d685884f1fb8a1e08b9c9eb2370e1ff77569a7e"} Dec 05 10:59:17.386559 master-0 kubenswrapper[24928]: I1205 10:59:17.384234 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz" event={"ID":"0c2e6173-1e2a-48ef-94ae-70974b457589","Type":"ContainerStarted","Data":"c0f5cf587c1ead4f27461d67c16c7b914a108277ff56de622f962a46cee38b8d"} Dec 05 10:59:17.386559 master-0 kubenswrapper[24928]: I1205 10:59:17.386232 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-nj5nk" event={"ID":"0962dd5b-ecea-4032-84e5-d848b7e46451","Type":"ContainerStarted","Data":"a27e8f74b419774b99e570ca0e40bc5e247a880637d3aaa6945295147420cf98"} Dec 05 10:59:17.411448 master-0 kubenswrapper[24928]: I1205 10:59:17.408540 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-4l8x5" podStartSLOduration=6.408514905 podStartE2EDuration="6.408514905s" podCreationTimestamp="2025-12-05 10:59:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:59:17.401259608 +0000 UTC m=+717.404453459" watchObservedRunningTime="2025-12-05 10:59:17.408514905 +0000 UTC m=+717.411708756" Dec 05 10:59:20.423912 master-0 kubenswrapper[24928]: I1205 10:59:20.423327 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz" event={"ID":"0c2e6173-1e2a-48ef-94ae-70974b457589","Type":"ContainerStarted","Data":"e0bebcca81f95b2a4cac21df4bc93c8214d557a1daa05af66c28122d8424f142"} Dec 05 10:59:20.427332 master-0 kubenswrapper[24928]: I1205 10:59:20.427278 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-nj5nk" event={"ID":"0962dd5b-ecea-4032-84e5-d848b7e46451","Type":"ContainerStarted","Data":"428ed4afe865df17f43297375cc50add58b65c4248bd3e865c38076bc95b8ced"} Dec 05 10:59:20.430330 master-0 kubenswrapper[24928]: I1205 10:59:20.429986 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29" event={"ID":"686c75b7-f169-41df-aa2e-80f48a404f30","Type":"ContainerStarted","Data":"7a8a00279f0428859214896092c1d956ebf89be33a69e99f661ecd94f156d85e"} Dec 05 10:59:20.444834 master-0 kubenswrapper[24928]: I1205 10:59:20.444039 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-l52lz" podStartSLOduration=5.576673935 podStartE2EDuration="8.444019212s" podCreationTimestamp="2025-12-05 10:59:12 +0000 UTC" firstStartedPulling="2025-12-05 10:59:16.7302269 +0000 UTC m=+716.733420751" lastFinishedPulling="2025-12-05 10:59:19.597572137 +0000 UTC m=+719.600766028" observedRunningTime="2025-12-05 10:59:20.443023157 +0000 UTC m=+720.446217018" watchObservedRunningTime="2025-12-05 10:59:20.444019212 +0000 UTC m=+720.447213083" Dec 05 10:59:20.489448 master-0 kubenswrapper[24928]: I1205 10:59:20.482360 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-78b56678b9-zpw29" podStartSLOduration=6.698220801 podStartE2EDuration="9.48233614s" podCreationTimestamp="2025-12-05 10:59:11 +0000 UTC" firstStartedPulling="2025-12-05 10:59:16.817139799 +0000 UTC m=+716.820333650" lastFinishedPulling="2025-12-05 10:59:19.601255138 +0000 UTC m=+719.604448989" observedRunningTime="2025-12-05 10:59:20.472239603 +0000 UTC m=+720.475433464" watchObservedRunningTime="2025-12-05 10:59:20.48233614 +0000 UTC m=+720.485529991" Dec 05 10:59:20.516452 master-0 kubenswrapper[24928]: I1205 10:59:20.516206 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-nj5nk" podStartSLOduration=6.314595226 podStartE2EDuration="9.51617079s" podCreationTimestamp="2025-12-05 10:59:11 +0000 UTC" firstStartedPulling="2025-12-05 10:59:16.397177752 +0000 UTC m=+716.400371603" lastFinishedPulling="2025-12-05 10:59:19.598753316 +0000 UTC m=+719.601947167" observedRunningTime="2025-12-05 10:59:20.502061793 +0000 UTC m=+720.505255664" watchObservedRunningTime="2025-12-05 10:59:20.51617079 +0000 UTC m=+720.519364651" Dec 05 10:59:22.458556 master-0 kubenswrapper[24928]: I1205 10:59:22.458452 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-jw8mn" event={"ID":"7cd9e16b-26c6-4b23-bbbd-a9e4c7cdd0b6","Type":"ContainerStarted","Data":"15fe3a4e3c133fc330f44bbd95b29c93433a75c416a5d876c4f83d951c25825e"} Dec 05 10:59:22.459213 master-0 kubenswrapper[24928]: I1205 10:59:22.458820 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-jw8mn" Dec 05 10:59:22.497955 master-0 kubenswrapper[24928]: I1205 10:59:22.497864 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-jw8mn" podStartSLOduration=3.18070953 podStartE2EDuration="7.497845902s" podCreationTimestamp="2025-12-05 10:59:15 +0000 UTC" firstStartedPulling="2025-12-05 10:59:16.976182995 +0000 UTC m=+716.979376846" lastFinishedPulling="2025-12-05 10:59:21.293319367 +0000 UTC m=+721.296513218" observedRunningTime="2025-12-05 10:59:22.488504243 +0000 UTC m=+722.491698114" watchObservedRunningTime="2025-12-05 10:59:22.497845902 +0000 UTC m=+722.501039753" Dec 05 10:59:25.073549 master-0 kubenswrapper[24928]: I1205 10:59:25.073486 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5f77dd7bb4-xmg4x" Dec 05 10:59:25.485512 master-0 kubenswrapper[24928]: I1205 10:59:25.485310 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-g242x" event={"ID":"4f87e93b-6e0f-4ecf-80f7-46c3fd613ae6","Type":"ContainerStarted","Data":"081ce8b6acbe90c57fe2e5bebb41169f9ced625f5d88f1f21b39593c0e4841b5"} Dec 05 10:59:25.485818 master-0 kubenswrapper[24928]: I1205 10:59:25.485687 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-g242x" Dec 05 10:59:25.488046 master-0 kubenswrapper[24928]: I1205 10:59:25.488010 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-g242x" Dec 05 10:59:25.511013 master-0 kubenswrapper[24928]: I1205 10:59:25.510914 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-g242x" podStartSLOduration=2.763130722 podStartE2EDuration="10.51089518s" podCreationTimestamp="2025-12-05 10:59:15 +0000 UTC" firstStartedPulling="2025-12-05 10:59:16.884385416 +0000 UTC m=+716.887579267" lastFinishedPulling="2025-12-05 10:59:24.632149864 +0000 UTC m=+724.635343725" observedRunningTime="2025-12-05 10:59:25.50643592 +0000 UTC m=+725.509629771" watchObservedRunningTime="2025-12-05 10:59:25.51089518 +0000 UTC m=+725.514089031" Dec 05 10:59:26.287875 master-0 kubenswrapper[24928]: I1205 10:59:26.287813 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-jw8mn" Dec 05 10:59:29.661467 master-0 kubenswrapper[24928]: I1205 10:59:29.661386 24928 scope.go:117] "RemoveContainer" containerID="5e1cd037da3a72a86565cde7e25ecd2da3f7f8cf8ea4a83328e0ab5b87b966ad" Dec 05 10:59:44.440849 master-0 kubenswrapper[24928]: I1205 10:59:44.440777 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-9d5bd9bc7-q878m" Dec 05 10:59:49.657192 master-0 kubenswrapper[24928]: I1205 10:59:49.657120 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb"] Dec 05 10:59:49.658542 master-0 kubenswrapper[24928]: I1205 10:59:49.658506 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" Dec 05 10:59:49.661125 master-0 kubenswrapper[24928]: I1205 10:59:49.661090 24928 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 05 10:59:49.668571 master-0 kubenswrapper[24928]: I1205 10:59:49.668536 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-2cn6b"] Dec 05 10:59:49.672259 master-0 kubenswrapper[24928]: I1205 10:59:49.672224 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.674239 master-0 kubenswrapper[24928]: I1205 10:59:49.674174 24928 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 05 10:59:49.680879 master-0 kubenswrapper[24928]: I1205 10:59:49.680827 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 05 10:59:49.692536 master-0 kubenswrapper[24928]: I1205 10:59:49.689562 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb"] Dec 05 10:59:49.796392 master-0 kubenswrapper[24928]: I1205 10:59:49.794031 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-9stls"] Dec 05 10:59:49.803455 master-0 kubenswrapper[24928]: I1205 10:59:49.801010 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9stls" Dec 05 10:59:49.803455 master-0 kubenswrapper[24928]: I1205 10:59:49.802731 24928 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 05 10:59:49.803455 master-0 kubenswrapper[24928]: I1205 10:59:49.802979 24928 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 05 10:59:49.803780 master-0 kubenswrapper[24928]: I1205 10:59:49.803593 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 05 10:59:49.811382 master-0 kubenswrapper[24928]: I1205 10:59:49.811320 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-fpl59"] Dec 05 10:59:49.815448 master-0 kubenswrapper[24928]: I1205 10:59:49.812952 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:49.815892 master-0 kubenswrapper[24928]: I1205 10:59:49.815846 24928 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 05 10:59:49.829506 master-0 kubenswrapper[24928]: I1205 10:59:49.828708 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w95xd\" (UniqueName: \"kubernetes.io/projected/fec8fc95-cfe1-4f43-b693-7a11680da86a-kube-api-access-w95xd\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.829506 master-0 kubenswrapper[24928]: I1205 10:59:49.828762 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnkmq\" (UniqueName: \"kubernetes.io/projected/a65480f6-aa9d-45b5-b3ba-541e87b7b145-kube-api-access-lnkmq\") pod \"frr-k8s-webhook-server-7fcb986d4-dlsnb\" (UID: \"a65480f6-aa9d-45b5-b3ba-541e87b7b145\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" Dec 05 10:59:49.829506 master-0 kubenswrapper[24928]: I1205 10:59:49.828799 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a65480f6-aa9d-45b5-b3ba-541e87b7b145-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-dlsnb\" (UID: \"a65480f6-aa9d-45b5-b3ba-541e87b7b145\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" Dec 05 10:59:49.829506 master-0 kubenswrapper[24928]: I1205 10:59:49.828837 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-frr-conf\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.829506 master-0 kubenswrapper[24928]: I1205 10:59:49.828878 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-frr-sockets\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.829506 master-0 kubenswrapper[24928]: I1205 10:59:49.828902 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-reloader\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.829506 master-0 kubenswrapper[24928]: I1205 10:59:49.828950 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-metrics\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.829506 master-0 kubenswrapper[24928]: I1205 10:59:49.828983 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fec8fc95-cfe1-4f43-b693-7a11680da86a-metrics-certs\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.829506 master-0 kubenswrapper[24928]: I1205 10:59:49.829006 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fec8fc95-cfe1-4f43-b693-7a11680da86a-frr-startup\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.844451 master-0 kubenswrapper[24928]: I1205 10:59:49.844055 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-fpl59"] Dec 05 10:59:49.939713 master-0 kubenswrapper[24928]: I1205 10:59:49.939572 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-frr-sockets\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.939969 master-0 kubenswrapper[24928]: I1205 10:59:49.939948 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-reloader\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.940080 master-0 kubenswrapper[24928]: I1205 10:59:49.940064 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a019084-2c01-4311-81d4-df4e5fe47107-cert\") pod \"controller-f8648f98b-fpl59\" (UID: \"2a019084-2c01-4311-81d4-df4e5fe47107\") " pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:49.940194 master-0 kubenswrapper[24928]: I1205 10:59:49.940179 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-metrics\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.940293 master-0 kubenswrapper[24928]: I1205 10:59:49.940278 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-metrics-certs\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:49.940397 master-0 kubenswrapper[24928]: I1205 10:59:49.940382 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-memberlist\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:49.940649 master-0 kubenswrapper[24928]: I1205 10:59:49.940577 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld72d\" (UniqueName: \"kubernetes.io/projected/b953bb78-052f-4b26-94d3-da9bab7a9fdc-kube-api-access-ld72d\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:49.940720 master-0 kubenswrapper[24928]: I1205 10:59:49.940680 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fec8fc95-cfe1-4f43-b693-7a11680da86a-metrics-certs\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.940769 master-0 kubenswrapper[24928]: I1205 10:59:49.940717 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b953bb78-052f-4b26-94d3-da9bab7a9fdc-metallb-excludel2\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:49.940769 master-0 kubenswrapper[24928]: I1205 10:59:49.940746 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fec8fc95-cfe1-4f43-b693-7a11680da86a-frr-startup\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.940854 master-0 kubenswrapper[24928]: I1205 10:59:49.940817 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nh5m\" (UniqueName: \"kubernetes.io/projected/2a019084-2c01-4311-81d4-df4e5fe47107-kube-api-access-9nh5m\") pod \"controller-f8648f98b-fpl59\" (UID: \"2a019084-2c01-4311-81d4-df4e5fe47107\") " pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:49.940921 master-0 kubenswrapper[24928]: I1205 10:59:49.940898 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w95xd\" (UniqueName: \"kubernetes.io/projected/fec8fc95-cfe1-4f43-b693-7a11680da86a-kube-api-access-w95xd\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.940983 master-0 kubenswrapper[24928]: I1205 10:59:49.940931 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnkmq\" (UniqueName: \"kubernetes.io/projected/a65480f6-aa9d-45b5-b3ba-541e87b7b145-kube-api-access-lnkmq\") pod \"frr-k8s-webhook-server-7fcb986d4-dlsnb\" (UID: \"a65480f6-aa9d-45b5-b3ba-541e87b7b145\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" Dec 05 10:59:49.940983 master-0 kubenswrapper[24928]: I1205 10:59:49.940967 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a65480f6-aa9d-45b5-b3ba-541e87b7b145-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-dlsnb\" (UID: \"a65480f6-aa9d-45b5-b3ba-541e87b7b145\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" Dec 05 10:59:49.941065 master-0 kubenswrapper[24928]: I1205 10:59:49.941002 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a019084-2c01-4311-81d4-df4e5fe47107-metrics-certs\") pod \"controller-f8648f98b-fpl59\" (UID: \"2a019084-2c01-4311-81d4-df4e5fe47107\") " pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:49.941065 master-0 kubenswrapper[24928]: I1205 10:59:49.941044 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-frr-conf\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.941567 master-0 kubenswrapper[24928]: I1205 10:59:49.941546 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-metrics\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.945444 master-0 kubenswrapper[24928]: I1205 10:59:49.942535 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fec8fc95-cfe1-4f43-b693-7a11680da86a-frr-startup\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.945444 master-0 kubenswrapper[24928]: I1205 10:59:49.943839 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-frr-conf\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.945840 master-0 kubenswrapper[24928]: I1205 10:59:49.945818 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-frr-sockets\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.946974 master-0 kubenswrapper[24928]: I1205 10:59:49.946915 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fec8fc95-cfe1-4f43-b693-7a11680da86a-metrics-certs\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.955463 master-0 kubenswrapper[24928]: I1205 10:59:49.949849 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fec8fc95-cfe1-4f43-b693-7a11680da86a-reloader\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.967451 master-0 kubenswrapper[24928]: I1205 10:59:49.962294 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a65480f6-aa9d-45b5-b3ba-541e87b7b145-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-dlsnb\" (UID: \"a65480f6-aa9d-45b5-b3ba-541e87b7b145\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" Dec 05 10:59:49.985449 master-0 kubenswrapper[24928]: I1205 10:59:49.985070 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w95xd\" (UniqueName: \"kubernetes.io/projected/fec8fc95-cfe1-4f43-b693-7a11680da86a-kube-api-access-w95xd\") pod \"frr-k8s-2cn6b\" (UID: \"fec8fc95-cfe1-4f43-b693-7a11680da86a\") " pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:49.991447 master-0 kubenswrapper[24928]: I1205 10:59:49.990293 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnkmq\" (UniqueName: \"kubernetes.io/projected/a65480f6-aa9d-45b5-b3ba-541e87b7b145-kube-api-access-lnkmq\") pod \"frr-k8s-webhook-server-7fcb986d4-dlsnb\" (UID: \"a65480f6-aa9d-45b5-b3ba-541e87b7b145\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" Dec 05 10:59:50.032514 master-0 kubenswrapper[24928]: I1205 10:59:50.024211 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" Dec 05 10:59:50.035897 master-0 kubenswrapper[24928]: I1205 10:59:50.035834 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-2cn6b" Dec 05 10:59:50.042499 master-0 kubenswrapper[24928]: I1205 10:59:50.042220 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a019084-2c01-4311-81d4-df4e5fe47107-metrics-certs\") pod \"controller-f8648f98b-fpl59\" (UID: \"2a019084-2c01-4311-81d4-df4e5fe47107\") " pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:50.042499 master-0 kubenswrapper[24928]: I1205 10:59:50.042295 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a019084-2c01-4311-81d4-df4e5fe47107-cert\") pod \"controller-f8648f98b-fpl59\" (UID: \"2a019084-2c01-4311-81d4-df4e5fe47107\") " pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:50.042499 master-0 kubenswrapper[24928]: I1205 10:59:50.042333 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-metrics-certs\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:50.042499 master-0 kubenswrapper[24928]: I1205 10:59:50.042353 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-memberlist\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:50.042499 master-0 kubenswrapper[24928]: I1205 10:59:50.042371 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld72d\" (UniqueName: \"kubernetes.io/projected/b953bb78-052f-4b26-94d3-da9bab7a9fdc-kube-api-access-ld72d\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:50.042499 master-0 kubenswrapper[24928]: I1205 10:59:50.042390 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b953bb78-052f-4b26-94d3-da9bab7a9fdc-metallb-excludel2\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:50.042499 master-0 kubenswrapper[24928]: I1205 10:59:50.042421 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nh5m\" (UniqueName: \"kubernetes.io/projected/2a019084-2c01-4311-81d4-df4e5fe47107-kube-api-access-9nh5m\") pod \"controller-f8648f98b-fpl59\" (UID: \"2a019084-2c01-4311-81d4-df4e5fe47107\") " pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:50.042718 master-0 kubenswrapper[24928]: E1205 10:59:50.042643 24928 secret.go:189] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 05 10:59:50.042718 master-0 kubenswrapper[24928]: E1205 10:59:50.042697 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-memberlist podName:b953bb78-052f-4b26-94d3-da9bab7a9fdc nodeName:}" failed. No retries permitted until 2025-12-05 10:59:50.542677624 +0000 UTC m=+750.545871475 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-memberlist") pod "speaker-9stls" (UID: "b953bb78-052f-4b26-94d3-da9bab7a9fdc") : secret "metallb-memberlist" not found Dec 05 10:59:50.043545 master-0 kubenswrapper[24928]: I1205 10:59:50.043520 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b953bb78-052f-4b26-94d3-da9bab7a9fdc-metallb-excludel2\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:50.047797 master-0 kubenswrapper[24928]: I1205 10:59:50.047734 24928 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 10:59:50.052596 master-0 kubenswrapper[24928]: I1205 10:59:50.052549 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-metrics-certs\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:50.064570 master-0 kubenswrapper[24928]: I1205 10:59:50.064530 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2a019084-2c01-4311-81d4-df4e5fe47107-cert\") pod \"controller-f8648f98b-fpl59\" (UID: \"2a019084-2c01-4311-81d4-df4e5fe47107\") " pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:50.070274 master-0 kubenswrapper[24928]: I1205 10:59:50.070204 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2a019084-2c01-4311-81d4-df4e5fe47107-metrics-certs\") pod \"controller-f8648f98b-fpl59\" (UID: \"2a019084-2c01-4311-81d4-df4e5fe47107\") " pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:50.086127 master-0 kubenswrapper[24928]: I1205 10:59:50.085843 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld72d\" (UniqueName: \"kubernetes.io/projected/b953bb78-052f-4b26-94d3-da9bab7a9fdc-kube-api-access-ld72d\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:50.086127 master-0 kubenswrapper[24928]: I1205 10:59:50.086026 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nh5m\" (UniqueName: \"kubernetes.io/projected/2a019084-2c01-4311-81d4-df4e5fe47107-kube-api-access-9nh5m\") pod \"controller-f8648f98b-fpl59\" (UID: \"2a019084-2c01-4311-81d4-df4e5fe47107\") " pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:50.157937 master-0 kubenswrapper[24928]: I1205 10:59:50.157892 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:50.339375 master-0 kubenswrapper[24928]: I1205 10:59:50.339118 24928 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:59:50.503166 master-0 kubenswrapper[24928]: I1205 10:59:50.503112 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb"] Dec 05 10:59:50.556564 master-0 kubenswrapper[24928]: I1205 10:59:50.556502 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-memberlist\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:50.556768 master-0 kubenswrapper[24928]: E1205 10:59:50.556699 24928 secret.go:189] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 05 10:59:50.556849 master-0 kubenswrapper[24928]: E1205 10:59:50.556776 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-memberlist podName:b953bb78-052f-4b26-94d3-da9bab7a9fdc nodeName:}" failed. No retries permitted until 2025-12-05 10:59:51.556757636 +0000 UTC m=+751.559951487 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-memberlist") pod "speaker-9stls" (UID: "b953bb78-052f-4b26-94d3-da9bab7a9fdc") : secret "metallb-memberlist" not found Dec 05 10:59:50.618220 master-0 kubenswrapper[24928]: I1205 10:59:50.618158 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-fpl59"] Dec 05 10:59:50.618626 master-0 kubenswrapper[24928]: W1205 10:59:50.618592 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a019084_2c01_4311_81d4_df4e5fe47107.slice/crio-589f3aee4ba52f2201828199fd23b811f5d433ceeb6ddcbb9fc4e7e17f9158ee WatchSource:0}: Error finding container 589f3aee4ba52f2201828199fd23b811f5d433ceeb6ddcbb9fc4e7e17f9158ee: Status 404 returned error can't find the container with id 589f3aee4ba52f2201828199fd23b811f5d433ceeb6ddcbb9fc4e7e17f9158ee Dec 05 10:59:50.716502 master-0 kubenswrapper[24928]: I1205 10:59:50.714208 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-fpl59" event={"ID":"2a019084-2c01-4311-81d4-df4e5fe47107","Type":"ContainerStarted","Data":"589f3aee4ba52f2201828199fd23b811f5d433ceeb6ddcbb9fc4e7e17f9158ee"} Dec 05 10:59:50.716502 master-0 kubenswrapper[24928]: I1205 10:59:50.715923 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" event={"ID":"a65480f6-aa9d-45b5-b3ba-541e87b7b145","Type":"ContainerStarted","Data":"f5eb71fd3431cd6a95f422c6da578757722a2daeedbc3247db824aebf2e71d1e"} Dec 05 10:59:50.717344 master-0 kubenswrapper[24928]: I1205 10:59:50.717271 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cn6b" event={"ID":"fec8fc95-cfe1-4f43-b693-7a11680da86a","Type":"ContainerStarted","Data":"c9ab638e97d22ef55955f101295221b111fafa5e460571da7e28924e2b0fdf75"} Dec 05 10:59:51.582043 master-0 kubenswrapper[24928]: I1205 10:59:51.581969 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-memberlist\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:51.586256 master-0 kubenswrapper[24928]: I1205 10:59:51.586200 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b953bb78-052f-4b26-94d3-da9bab7a9fdc-memberlist\") pod \"speaker-9stls\" (UID: \"b953bb78-052f-4b26-94d3-da9bab7a9fdc\") " pod="metallb-system/speaker-9stls" Dec 05 10:59:51.639473 master-0 kubenswrapper[24928]: I1205 10:59:51.638551 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-9stls" Dec 05 10:59:51.733454 master-0 kubenswrapper[24928]: I1205 10:59:51.731733 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9stls" event={"ID":"b953bb78-052f-4b26-94d3-da9bab7a9fdc","Type":"ContainerStarted","Data":"c7a49e6b6504a10f3cfd8cc6d7eeb27b73bc3cd26595abce3c1c44279c06c94e"} Dec 05 10:59:51.737449 master-0 kubenswrapper[24928]: I1205 10:59:51.734336 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-fpl59" event={"ID":"2a019084-2c01-4311-81d4-df4e5fe47107","Type":"ContainerStarted","Data":"150226e1dce6cb929ae19876225b0413487696bf59fd13e1cc7b1cfec243dc00"} Dec 05 10:59:51.972090 master-0 kubenswrapper[24928]: I1205 10:59:51.970577 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk"] Dec 05 10:59:51.973220 master-0 kubenswrapper[24928]: I1205 10:59:51.973178 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk" Dec 05 10:59:51.992488 master-0 kubenswrapper[24928]: I1205 10:59:51.992443 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb"] Dec 05 10:59:51.993721 master-0 kubenswrapper[24928]: I1205 10:59:51.993693 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" Dec 05 10:59:51.996551 master-0 kubenswrapper[24928]: I1205 10:59:51.996164 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 05 10:59:52.010113 master-0 kubenswrapper[24928]: I1205 10:59:52.008472 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-hxkln"] Dec 05 10:59:52.013341 master-0 kubenswrapper[24928]: I1205 10:59:52.013305 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.018092 master-0 kubenswrapper[24928]: I1205 10:59:52.018037 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb"] Dec 05 10:59:52.028190 master-0 kubenswrapper[24928]: I1205 10:59:52.025551 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk"] Dec 05 10:59:52.096565 master-0 kubenswrapper[24928]: I1205 10:59:52.092791 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwq9p\" (UniqueName: \"kubernetes.io/projected/4c719724-d364-4e6b-8c86-b72d3152536d-kube-api-access-dwq9p\") pod \"nmstate-webhook-5f6d4c5ccb-jxlrb\" (UID: \"4c719724-d364-4e6b-8c86-b72d3152536d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" Dec 05 10:59:52.096565 master-0 kubenswrapper[24928]: I1205 10:59:52.092859 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c719724-d364-4e6b-8c86-b72d3152536d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jxlrb\" (UID: \"4c719724-d364-4e6b-8c86-b72d3152536d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" Dec 05 10:59:52.096565 master-0 kubenswrapper[24928]: I1205 10:59:52.092935 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stfsr\" (UniqueName: \"kubernetes.io/projected/1501ec1c-63ed-41c9-8751-affef02bd884-kube-api-access-stfsr\") pod \"nmstate-metrics-7f946cbc9-jqhpk\" (UID: \"1501ec1c-63ed-41c9-8751-affef02bd884\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk" Dec 05 10:59:52.151720 master-0 kubenswrapper[24928]: I1205 10:59:52.145657 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn"] Dec 05 10:59:52.151720 master-0 kubenswrapper[24928]: I1205 10:59:52.147466 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:52.154094 master-0 kubenswrapper[24928]: I1205 10:59:52.154060 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 05 10:59:52.154266 master-0 kubenswrapper[24928]: I1205 10:59:52.154152 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 05 10:59:52.157623 master-0 kubenswrapper[24928]: I1205 10:59:52.157594 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn"] Dec 05 10:59:52.195266 master-0 kubenswrapper[24928]: I1205 10:59:52.195201 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stfsr\" (UniqueName: \"kubernetes.io/projected/1501ec1c-63ed-41c9-8751-affef02bd884-kube-api-access-stfsr\") pod \"nmstate-metrics-7f946cbc9-jqhpk\" (UID: \"1501ec1c-63ed-41c9-8751-affef02bd884\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk" Dec 05 10:59:52.195266 master-0 kubenswrapper[24928]: I1205 10:59:52.195265 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kgtn\" (UniqueName: \"kubernetes.io/projected/27c2f18f-89e8-4439-a383-6bc8f6bed3af-kube-api-access-9kgtn\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.197671 master-0 kubenswrapper[24928]: I1205 10:59:52.195305 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/27c2f18f-89e8-4439-a383-6bc8f6bed3af-nmstate-lock\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.197671 master-0 kubenswrapper[24928]: I1205 10:59:52.195380 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/27c2f18f-89e8-4439-a383-6bc8f6bed3af-dbus-socket\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.197671 master-0 kubenswrapper[24928]: I1205 10:59:52.195413 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwq9p\" (UniqueName: \"kubernetes.io/projected/4c719724-d364-4e6b-8c86-b72d3152536d-kube-api-access-dwq9p\") pod \"nmstate-webhook-5f6d4c5ccb-jxlrb\" (UID: \"4c719724-d364-4e6b-8c86-b72d3152536d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" Dec 05 10:59:52.197671 master-0 kubenswrapper[24928]: I1205 10:59:52.195770 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c719724-d364-4e6b-8c86-b72d3152536d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jxlrb\" (UID: \"4c719724-d364-4e6b-8c86-b72d3152536d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" Dec 05 10:59:52.197671 master-0 kubenswrapper[24928]: I1205 10:59:52.195847 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/27c2f18f-89e8-4439-a383-6bc8f6bed3af-ovs-socket\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.204467 master-0 kubenswrapper[24928]: I1205 10:59:52.203163 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/4c719724-d364-4e6b-8c86-b72d3152536d-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-jxlrb\" (UID: \"4c719724-d364-4e6b-8c86-b72d3152536d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" Dec 05 10:59:52.302905 master-0 kubenswrapper[24928]: I1205 10:59:52.302822 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/27c2f18f-89e8-4439-a383-6bc8f6bed3af-nmstate-lock\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.310578 master-0 kubenswrapper[24928]: I1205 10:59:52.302927 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/27c2f18f-89e8-4439-a383-6bc8f6bed3af-dbus-socket\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.310578 master-0 kubenswrapper[24928]: I1205 10:59:52.303012 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/27c2f18f-89e8-4439-a383-6bc8f6bed3af-dbus-socket\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.310578 master-0 kubenswrapper[24928]: I1205 10:59:52.303005 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/27c2f18f-89e8-4439-a383-6bc8f6bed3af-nmstate-lock\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.310578 master-0 kubenswrapper[24928]: I1205 10:59:52.303049 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:52.310578 master-0 kubenswrapper[24928]: I1205 10:59:52.303332 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhg95\" (UniqueName: \"kubernetes.io/projected/381f465d-c34b-46c0-99f6-d50a6f6307bb-kube-api-access-mhg95\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:52.310578 master-0 kubenswrapper[24928]: I1205 10:59:52.303408 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/381f465d-c34b-46c0-99f6-d50a6f6307bb-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:52.310578 master-0 kubenswrapper[24928]: I1205 10:59:52.303500 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/27c2f18f-89e8-4439-a383-6bc8f6bed3af-ovs-socket\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.310578 master-0 kubenswrapper[24928]: I1205 10:59:52.303684 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/27c2f18f-89e8-4439-a383-6bc8f6bed3af-ovs-socket\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.310578 master-0 kubenswrapper[24928]: I1205 10:59:52.303725 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kgtn\" (UniqueName: \"kubernetes.io/projected/27c2f18f-89e8-4439-a383-6bc8f6bed3af-kube-api-access-9kgtn\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:52.406190 master-0 kubenswrapper[24928]: I1205 10:59:52.405992 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhg95\" (UniqueName: \"kubernetes.io/projected/381f465d-c34b-46c0-99f6-d50a6f6307bb-kube-api-access-mhg95\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:52.406190 master-0 kubenswrapper[24928]: I1205 10:59:52.406062 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/381f465d-c34b-46c0-99f6-d50a6f6307bb-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:52.407235 master-0 kubenswrapper[24928]: I1205 10:59:52.407186 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/381f465d-c34b-46c0-99f6-d50a6f6307bb-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:52.407352 master-0 kubenswrapper[24928]: I1205 10:59:52.407275 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:52.407521 master-0 kubenswrapper[24928]: E1205 10:59:52.407322 24928 secret.go:189] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 05 10:59:52.407586 master-0 kubenswrapper[24928]: E1205 10:59:52.407565 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert podName:381f465d-c34b-46c0-99f6-d50a6f6307bb nodeName:}" failed. No retries permitted until 2025-12-05 10:59:52.907544413 +0000 UTC m=+752.910738474 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-hvbdn" (UID: "381f465d-c34b-46c0-99f6-d50a6f6307bb") : secret "plugin-serving-cert" not found Dec 05 10:59:52.745545 master-0 kubenswrapper[24928]: I1205 10:59:52.745404 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9stls" event={"ID":"b953bb78-052f-4b26-94d3-da9bab7a9fdc","Type":"ContainerStarted","Data":"52afaf1e371efd331ce4697d391ee3fa602be22c8f7b601a9a3a79a2d251eab2"} Dec 05 10:59:52.917280 master-0 kubenswrapper[24928]: I1205 10:59:52.917206 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:52.917665 master-0 kubenswrapper[24928]: E1205 10:59:52.917526 24928 secret.go:189] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 05 10:59:52.917793 master-0 kubenswrapper[24928]: E1205 10:59:52.917766 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert podName:381f465d-c34b-46c0-99f6-d50a6f6307bb nodeName:}" failed. No retries permitted until 2025-12-05 10:59:53.91773145 +0000 UTC m=+753.920925301 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-hvbdn" (UID: "381f465d-c34b-46c0-99f6-d50a6f6307bb") : secret "plugin-serving-cert" not found Dec 05 10:59:53.938605 master-0 kubenswrapper[24928]: I1205 10:59:53.938536 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:53.939081 master-0 kubenswrapper[24928]: E1205 10:59:53.938745 24928 secret.go:189] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 05 10:59:53.939081 master-0 kubenswrapper[24928]: E1205 10:59:53.938852 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert podName:381f465d-c34b-46c0-99f6-d50a6f6307bb nodeName:}" failed. No retries permitted until 2025-12-05 10:59:55.938826662 +0000 UTC m=+755.942020523 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-hvbdn" (UID: "381f465d-c34b-46c0-99f6-d50a6f6307bb") : secret "plugin-serving-cert" not found Dec 05 10:59:54.548621 master-0 kubenswrapper[24928]: I1205 10:59:54.538616 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stfsr\" (UniqueName: \"kubernetes.io/projected/1501ec1c-63ed-41c9-8751-affef02bd884-kube-api-access-stfsr\") pod \"nmstate-metrics-7f946cbc9-jqhpk\" (UID: \"1501ec1c-63ed-41c9-8751-affef02bd884\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk" Dec 05 10:59:54.548621 master-0 kubenswrapper[24928]: I1205 10:59:54.543516 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kgtn\" (UniqueName: \"kubernetes.io/projected/27c2f18f-89e8-4439-a383-6bc8f6bed3af-kube-api-access-9kgtn\") pod \"nmstate-handler-hxkln\" (UID: \"27c2f18f-89e8-4439-a383-6bc8f6bed3af\") " pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:54.548621 master-0 kubenswrapper[24928]: I1205 10:59:54.546553 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwq9p\" (UniqueName: \"kubernetes.io/projected/4c719724-d364-4e6b-8c86-b72d3152536d-kube-api-access-dwq9p\") pod \"nmstate-webhook-5f6d4c5ccb-jxlrb\" (UID: \"4c719724-d364-4e6b-8c86-b72d3152536d\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" Dec 05 10:59:54.555569 master-0 kubenswrapper[24928]: I1205 10:59:54.555502 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhg95\" (UniqueName: \"kubernetes.io/projected/381f465d-c34b-46c0-99f6-d50a6f6307bb-kube-api-access-mhg95\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:54.646828 master-0 kubenswrapper[24928]: I1205 10:59:54.646750 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-78d584df9-x54pl"] Dec 05 10:59:54.656565 master-0 kubenswrapper[24928]: I1205 10:59:54.652184 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.664367 master-0 kubenswrapper[24928]: I1205 10:59:54.664308 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-78d584df9-x54pl"] Dec 05 10:59:54.712498 master-0 kubenswrapper[24928]: I1205 10:59:54.712395 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk" Dec 05 10:59:54.733311 master-0 kubenswrapper[24928]: I1205 10:59:54.729770 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" Dec 05 10:59:54.740976 master-0 kubenswrapper[24928]: I1205 10:59:54.740905 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 10:59:54.758436 master-0 kubenswrapper[24928]: I1205 10:59:54.758356 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-console-config\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.758620 master-0 kubenswrapper[24928]: I1205 10:59:54.758540 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e52dfad5-4a11-479b-8e34-276840f26282-console-serving-cert\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.759651 master-0 kubenswrapper[24928]: I1205 10:59:54.759610 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-service-ca\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.759651 master-0 kubenswrapper[24928]: I1205 10:59:54.759661 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-trusted-ca-bundle\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.759651 master-0 kubenswrapper[24928]: I1205 10:59:54.759734 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e52dfad5-4a11-479b-8e34-276840f26282-console-oauth-config\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.759651 master-0 kubenswrapper[24928]: I1205 10:59:54.759766 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-oauth-serving-cert\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.759651 master-0 kubenswrapper[24928]: I1205 10:59:54.759804 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwt29\" (UniqueName: \"kubernetes.io/projected/e52dfad5-4a11-479b-8e34-276840f26282-kube-api-access-rwt29\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.868019 master-0 kubenswrapper[24928]: I1205 10:59:54.862452 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e52dfad5-4a11-479b-8e34-276840f26282-console-serving-cert\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.868019 master-0 kubenswrapper[24928]: I1205 10:59:54.862884 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-service-ca\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.868019 master-0 kubenswrapper[24928]: I1205 10:59:54.863006 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-trusted-ca-bundle\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.868019 master-0 kubenswrapper[24928]: I1205 10:59:54.863164 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e52dfad5-4a11-479b-8e34-276840f26282-console-oauth-config\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.868019 master-0 kubenswrapper[24928]: I1205 10:59:54.863230 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-oauth-serving-cert\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.868019 master-0 kubenswrapper[24928]: I1205 10:59:54.863269 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwt29\" (UniqueName: \"kubernetes.io/projected/e52dfad5-4a11-479b-8e34-276840f26282-kube-api-access-rwt29\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.868019 master-0 kubenswrapper[24928]: I1205 10:59:54.863407 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-console-config\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.868019 master-0 kubenswrapper[24928]: I1205 10:59:54.867579 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e52dfad5-4a11-479b-8e34-276840f26282-console-serving-cert\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.868445 master-0 kubenswrapper[24928]: I1205 10:59:54.868255 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-console-config\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.869100 master-0 kubenswrapper[24928]: I1205 10:59:54.869050 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-service-ca\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.869341 master-0 kubenswrapper[24928]: I1205 10:59:54.869249 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-trusted-ca-bundle\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.873455 master-0 kubenswrapper[24928]: I1205 10:59:54.870699 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e52dfad5-4a11-479b-8e34-276840f26282-oauth-serving-cert\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.873455 master-0 kubenswrapper[24928]: W1205 10:59:54.871467 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27c2f18f_89e8_4439_a383_6bc8f6bed3af.slice/crio-7eacb31e980317023836ab34cf807af75de119b1aa6861aabb33f9caf1506f74 WatchSource:0}: Error finding container 7eacb31e980317023836ab34cf807af75de119b1aa6861aabb33f9caf1506f74: Status 404 returned error can't find the container with id 7eacb31e980317023836ab34cf807af75de119b1aa6861aabb33f9caf1506f74 Dec 05 10:59:54.873455 master-0 kubenswrapper[24928]: I1205 10:59:54.871990 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e52dfad5-4a11-479b-8e34-276840f26282-console-oauth-config\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:54.900638 master-0 kubenswrapper[24928]: I1205 10:59:54.890234 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwt29\" (UniqueName: \"kubernetes.io/projected/e52dfad5-4a11-479b-8e34-276840f26282-kube-api-access-rwt29\") pod \"console-78d584df9-x54pl\" (UID: \"e52dfad5-4a11-479b-8e34-276840f26282\") " pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:55.119573 master-0 kubenswrapper[24928]: I1205 10:59:55.107811 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-78d584df9-x54pl" Dec 05 10:59:55.184161 master-0 kubenswrapper[24928]: I1205 10:59:55.184094 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk"] Dec 05 10:59:55.459657 master-0 kubenswrapper[24928]: I1205 10:59:55.459596 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb"] Dec 05 10:59:55.790673 master-0 kubenswrapper[24928]: I1205 10:59:55.790615 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-9stls" event={"ID":"b953bb78-052f-4b26-94d3-da9bab7a9fdc","Type":"ContainerStarted","Data":"d2f59b10ac20bd56edd9ef789c6959917e8817d475cd03b9253d372214b0b393"} Dec 05 10:59:55.790894 master-0 kubenswrapper[24928]: I1205 10:59:55.790701 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-9stls" Dec 05 10:59:55.793232 master-0 kubenswrapper[24928]: I1205 10:59:55.793206 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-fpl59" event={"ID":"2a019084-2c01-4311-81d4-df4e5fe47107","Type":"ContainerStarted","Data":"2f2933ee9d7e2a35d440eb8d19efae717c4c10a610abff5ff03b6aeae7a1296c"} Dec 05 10:59:55.793344 master-0 kubenswrapper[24928]: I1205 10:59:55.793326 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 10:59:55.794880 master-0 kubenswrapper[24928]: I1205 10:59:55.794797 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hxkln" event={"ID":"27c2f18f-89e8-4439-a383-6bc8f6bed3af","Type":"ContainerStarted","Data":"7eacb31e980317023836ab34cf807af75de119b1aa6861aabb33f9caf1506f74"} Dec 05 10:59:55.818653 master-0 kubenswrapper[24928]: I1205 10:59:55.818541 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-9stls" podStartSLOduration=4.234321335 podStartE2EDuration="6.818526677s" podCreationTimestamp="2025-12-05 10:59:49 +0000 UTC" firstStartedPulling="2025-12-05 10:59:52.01715505 +0000 UTC m=+752.020348901" lastFinishedPulling="2025-12-05 10:59:54.601360392 +0000 UTC m=+754.604554243" observedRunningTime="2025-12-05 10:59:55.817103923 +0000 UTC m=+755.820297794" watchObservedRunningTime="2025-12-05 10:59:55.818526677 +0000 UTC m=+755.821720528" Dec 05 10:59:55.846341 master-0 kubenswrapper[24928]: I1205 10:59:55.844977 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-fpl59" podStartSLOduration=3.013708456 podStartE2EDuration="6.844956795s" podCreationTimestamp="2025-12-05 10:59:49 +0000 UTC" firstStartedPulling="2025-12-05 10:59:50.752510422 +0000 UTC m=+750.755704273" lastFinishedPulling="2025-12-05 10:59:54.583758761 +0000 UTC m=+754.586952612" observedRunningTime="2025-12-05 10:59:55.83823828 +0000 UTC m=+755.841432121" watchObservedRunningTime="2025-12-05 10:59:55.844956795 +0000 UTC m=+755.848150656" Dec 05 10:59:55.990504 master-0 kubenswrapper[24928]: I1205 10:59:55.990376 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:55.995013 master-0 kubenswrapper[24928]: I1205 10:59:55.994943 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/381f465d-c34b-46c0-99f6-d50a6f6307bb-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-hvbdn\" (UID: \"381f465d-c34b-46c0-99f6-d50a6f6307bb\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:56.089273 master-0 kubenswrapper[24928]: I1205 10:59:56.089140 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" Dec 05 10:59:58.226951 master-0 kubenswrapper[24928]: W1205 10:59:58.226265 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1501ec1c_63ed_41c9_8751_affef02bd884.slice/crio-ed8cb633ed057280daafbb92d454c639ba947d007d19793ac04234c249f01737 WatchSource:0}: Error finding container ed8cb633ed057280daafbb92d454c639ba947d007d19793ac04234c249f01737: Status 404 returned error can't find the container with id ed8cb633ed057280daafbb92d454c639ba947d007d19793ac04234c249f01737 Dec 05 10:59:58.792803 master-0 kubenswrapper[24928]: I1205 10:59:58.792711 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn"] Dec 05 10:59:58.833003 master-0 kubenswrapper[24928]: I1205 10:59:58.832939 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" event={"ID":"381f465d-c34b-46c0-99f6-d50a6f6307bb","Type":"ContainerStarted","Data":"f70445de15715e2d4f159e52278d90a702fd84e9c3e13f9f0ead6e7576fea2e1"} Dec 05 10:59:58.835266 master-0 kubenswrapper[24928]: I1205 10:59:58.835182 24928 generic.go:334] "Generic (PLEG): container finished" podID="fec8fc95-cfe1-4f43-b693-7a11680da86a" containerID="1e9d6ceedafde1d986f57939843d5a2180aeb529b300ccc34f672921147afda0" exitCode=0 Dec 05 10:59:58.835266 master-0 kubenswrapper[24928]: I1205 10:59:58.835247 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cn6b" event={"ID":"fec8fc95-cfe1-4f43-b693-7a11680da86a","Type":"ContainerDied","Data":"1e9d6ceedafde1d986f57939843d5a2180aeb529b300ccc34f672921147afda0"} Dec 05 10:59:58.836832 master-0 kubenswrapper[24928]: I1205 10:59:58.836772 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk" event={"ID":"1501ec1c-63ed-41c9-8751-affef02bd884","Type":"ContainerStarted","Data":"ed8cb633ed057280daafbb92d454c639ba947d007d19793ac04234c249f01737"} Dec 05 10:59:58.838347 master-0 kubenswrapper[24928]: I1205 10:59:58.838307 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" event={"ID":"4c719724-d364-4e6b-8c86-b72d3152536d","Type":"ContainerStarted","Data":"f9ff92119d2c04bf052d0fbca7b55f2ad907588dd4f28005c18810f9f0baf779"} Dec 05 10:59:58.840121 master-0 kubenswrapper[24928]: I1205 10:59:58.840007 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" event={"ID":"a65480f6-aa9d-45b5-b3ba-541e87b7b145","Type":"ContainerStarted","Data":"df912e5e253ed5f26dfef266369a739134c1cafcb2f3b20d2fb3a4048888add0"} Dec 05 10:59:58.840295 master-0 kubenswrapper[24928]: I1205 10:59:58.840193 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" Dec 05 10:59:58.889523 master-0 kubenswrapper[24928]: I1205 10:59:58.889454 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-78d584df9-x54pl"] Dec 05 10:59:58.890347 master-0 kubenswrapper[24928]: I1205 10:59:58.890246 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" podStartSLOduration=2.01331415 podStartE2EDuration="9.890230841s" podCreationTimestamp="2025-12-05 10:59:49 +0000 UTC" firstStartedPulling="2025-12-05 10:59:50.509288064 +0000 UTC m=+750.512481915" lastFinishedPulling="2025-12-05 10:59:58.386204755 +0000 UTC m=+758.389398606" observedRunningTime="2025-12-05 10:59:58.885485034 +0000 UTC m=+758.888678895" watchObservedRunningTime="2025-12-05 10:59:58.890230841 +0000 UTC m=+758.893424692" Dec 05 10:59:58.895830 master-0 kubenswrapper[24928]: W1205 10:59:58.895774 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode52dfad5_4a11_479b_8e34_276840f26282.slice/crio-4d62145e78731db529295ab5d9ba5ccbe6f0341e4f87f6fd5968338ff2727f5c WatchSource:0}: Error finding container 4d62145e78731db529295ab5d9ba5ccbe6f0341e4f87f6fd5968338ff2727f5c: Status 404 returned error can't find the container with id 4d62145e78731db529295ab5d9ba5ccbe6f0341e4f87f6fd5968338ff2727f5c Dec 05 10:59:59.851051 master-0 kubenswrapper[24928]: I1205 10:59:59.850824 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78d584df9-x54pl" event={"ID":"e52dfad5-4a11-479b-8e34-276840f26282","Type":"ContainerStarted","Data":"e18165d8d3ef4b086aabc0567348a5f157725bc24e05fb9ae1e7cbd248fdf747"} Dec 05 10:59:59.851051 master-0 kubenswrapper[24928]: I1205 10:59:59.850881 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-78d584df9-x54pl" event={"ID":"e52dfad5-4a11-479b-8e34-276840f26282","Type":"ContainerStarted","Data":"4d62145e78731db529295ab5d9ba5ccbe6f0341e4f87f6fd5968338ff2727f5c"} Dec 05 10:59:59.854537 master-0 kubenswrapper[24928]: I1205 10:59:59.854477 24928 generic.go:334] "Generic (PLEG): container finished" podID="fec8fc95-cfe1-4f43-b693-7a11680da86a" containerID="685326e06d9ddd4d95fd987267032d62f7ec22e7c08f4d88fb2ecf8359897f99" exitCode=0 Dec 05 10:59:59.854723 master-0 kubenswrapper[24928]: I1205 10:59:59.854595 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cn6b" event={"ID":"fec8fc95-cfe1-4f43-b693-7a11680da86a","Type":"ContainerDied","Data":"685326e06d9ddd4d95fd987267032d62f7ec22e7c08f4d88fb2ecf8359897f99"} Dec 05 10:59:59.905302 master-0 kubenswrapper[24928]: I1205 10:59:59.905205 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-78d584df9-x54pl" podStartSLOduration=5.905185153 podStartE2EDuration="5.905185153s" podCreationTimestamp="2025-12-05 10:59:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:59:59.873152979 +0000 UTC m=+759.876346840" watchObservedRunningTime="2025-12-05 10:59:59.905185153 +0000 UTC m=+759.908379004" Dec 05 11:00:00.167274 master-0 kubenswrapper[24928]: I1205 11:00:00.167223 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-fpl59" Dec 05 11:00:00.177213 master-0 kubenswrapper[24928]: I1205 11:00:00.177169 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm"] Dec 05 11:00:00.184289 master-0 kubenswrapper[24928]: I1205 11:00:00.181483 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:00.185403 master-0 kubenswrapper[24928]: I1205 11:00:00.185355 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-fdxl6" Dec 05 11:00:00.186632 master-0 kubenswrapper[24928]: I1205 11:00:00.186597 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 11:00:00.189269 master-0 kubenswrapper[24928]: I1205 11:00:00.189199 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm"] Dec 05 11:00:00.264527 master-0 kubenswrapper[24928]: I1205 11:00:00.263034 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06b902d6-85f1-4661-8cc6-be82f74a966d-config-volume\") pod \"collect-profiles-29415540-dgqvm\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:00.264527 master-0 kubenswrapper[24928]: I1205 11:00:00.263174 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06b902d6-85f1-4661-8cc6-be82f74a966d-secret-volume\") pod \"collect-profiles-29415540-dgqvm\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:00.264527 master-0 kubenswrapper[24928]: I1205 11:00:00.263352 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmk4g\" (UniqueName: \"kubernetes.io/projected/06b902d6-85f1-4661-8cc6-be82f74a966d-kube-api-access-gmk4g\") pod \"collect-profiles-29415540-dgqvm\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:00.368621 master-0 kubenswrapper[24928]: I1205 11:00:00.367077 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06b902d6-85f1-4661-8cc6-be82f74a966d-secret-volume\") pod \"collect-profiles-29415540-dgqvm\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:00.368621 master-0 kubenswrapper[24928]: I1205 11:00:00.367190 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmk4g\" (UniqueName: \"kubernetes.io/projected/06b902d6-85f1-4661-8cc6-be82f74a966d-kube-api-access-gmk4g\") pod \"collect-profiles-29415540-dgqvm\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:00.368621 master-0 kubenswrapper[24928]: I1205 11:00:00.367271 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06b902d6-85f1-4661-8cc6-be82f74a966d-config-volume\") pod \"collect-profiles-29415540-dgqvm\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:00.368621 master-0 kubenswrapper[24928]: I1205 11:00:00.368217 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06b902d6-85f1-4661-8cc6-be82f74a966d-config-volume\") pod \"collect-profiles-29415540-dgqvm\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:00.372095 master-0 kubenswrapper[24928]: I1205 11:00:00.372058 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06b902d6-85f1-4661-8cc6-be82f74a966d-secret-volume\") pod \"collect-profiles-29415540-dgqvm\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:00.385838 master-0 kubenswrapper[24928]: I1205 11:00:00.385786 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmk4g\" (UniqueName: \"kubernetes.io/projected/06b902d6-85f1-4661-8cc6-be82f74a966d-kube-api-access-gmk4g\") pod \"collect-profiles-29415540-dgqvm\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:00.509136 master-0 kubenswrapper[24928]: I1205 11:00:00.509050 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:00.865453 master-0 kubenswrapper[24928]: I1205 11:00:00.864878 24928 generic.go:334] "Generic (PLEG): container finished" podID="fec8fc95-cfe1-4f43-b693-7a11680da86a" containerID="daab493ee1e364d283f126de5866b331db417d55ad5e6889de475337a0457fbe" exitCode=0 Dec 05 11:00:00.865453 master-0 kubenswrapper[24928]: I1205 11:00:00.864932 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cn6b" event={"ID":"fec8fc95-cfe1-4f43-b693-7a11680da86a","Type":"ContainerDied","Data":"daab493ee1e364d283f126de5866b331db417d55ad5e6889de475337a0457fbe"} Dec 05 11:00:00.867746 master-0 kubenswrapper[24928]: I1205 11:00:00.867681 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-hxkln" event={"ID":"27c2f18f-89e8-4439-a383-6bc8f6bed3af","Type":"ContainerStarted","Data":"f064677e378a961552b45639a705d7fb46cb432bf6a151875dd8ff86cbda20c4"} Dec 05 11:00:00.867882 master-0 kubenswrapper[24928]: I1205 11:00:00.867856 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 11:00:00.870229 master-0 kubenswrapper[24928]: I1205 11:00:00.870173 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk" event={"ID":"1501ec1c-63ed-41c9-8751-affef02bd884","Type":"ContainerStarted","Data":"575262fe57df0d978cabc81ee20e9b26e4a5762ff974540ee674de291ac8431e"} Dec 05 11:00:00.870288 master-0 kubenswrapper[24928]: I1205 11:00:00.870233 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk" event={"ID":"1501ec1c-63ed-41c9-8751-affef02bd884","Type":"ContainerStarted","Data":"63bc49074c92033b0674822cb7b82306c2e4549e5838e36c08d58652a10e0157"} Dec 05 11:00:00.872170 master-0 kubenswrapper[24928]: I1205 11:00:00.872135 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" event={"ID":"4c719724-d364-4e6b-8c86-b72d3152536d","Type":"ContainerStarted","Data":"45640031f602bbbcd41aaa304330b371f5d1108f57262ccf24dc3b9b7ab44590"} Dec 05 11:00:00.914881 master-0 kubenswrapper[24928]: I1205 11:00:00.914790 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-hxkln" podStartSLOduration=4.848594284 podStartE2EDuration="9.914765384s" podCreationTimestamp="2025-12-05 10:59:51 +0000 UTC" firstStartedPulling="2025-12-05 10:59:54.88990128 +0000 UTC m=+754.893095131" lastFinishedPulling="2025-12-05 10:59:59.95607238 +0000 UTC m=+759.959266231" observedRunningTime="2025-12-05 11:00:00.914411615 +0000 UTC m=+760.917605466" watchObservedRunningTime="2025-12-05 11:00:00.914765384 +0000 UTC m=+760.917959245" Dec 05 11:00:00.942259 master-0 kubenswrapper[24928]: I1205 11:00:00.942102 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-jqhpk" podStartSLOduration=8.213301575 podStartE2EDuration="9.942081142s" podCreationTimestamp="2025-12-05 10:59:51 +0000 UTC" firstStartedPulling="2025-12-05 10:59:58.276820725 +0000 UTC m=+758.280014576" lastFinishedPulling="2025-12-05 11:00:00.005600292 +0000 UTC m=+760.008794143" observedRunningTime="2025-12-05 11:00:00.930487839 +0000 UTC m=+760.933681710" watchObservedRunningTime="2025-12-05 11:00:00.942081142 +0000 UTC m=+760.945274993" Dec 05 11:00:00.958527 master-0 kubenswrapper[24928]: I1205 11:00:00.957889 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" podStartSLOduration=8.240722538 podStartE2EDuration="9.95786443s" podCreationTimestamp="2025-12-05 10:59:51 +0000 UTC" firstStartedPulling="2025-12-05 10:59:58.276890247 +0000 UTC m=+758.280084098" lastFinishedPulling="2025-12-05 10:59:59.994032129 +0000 UTC m=+759.997225990" observedRunningTime="2025-12-05 11:00:00.953164804 +0000 UTC m=+760.956358665" watchObservedRunningTime="2025-12-05 11:00:00.95786443 +0000 UTC m=+760.961058281" Dec 05 11:00:01.475702 master-0 kubenswrapper[24928]: I1205 11:00:01.475498 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm"] Dec 05 11:00:01.642216 master-0 kubenswrapper[24928]: I1205 11:00:01.642146 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-9stls" Dec 05 11:00:01.888742 master-0 kubenswrapper[24928]: I1205 11:00:01.888654 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" event={"ID":"381f465d-c34b-46c0-99f6-d50a6f6307bb","Type":"ContainerStarted","Data":"cb832297e3289e9a6cd44d6b598e1baee45c3fd37d26cf8376edb78315e5bb67"} Dec 05 11:00:01.900676 master-0 kubenswrapper[24928]: I1205 11:00:01.900556 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cn6b" event={"ID":"fec8fc95-cfe1-4f43-b693-7a11680da86a","Type":"ContainerStarted","Data":"b85fc52056d27ca917a495c4e8e0aeee1efecfbd56c3224959794dd5eacbbd53"} Dec 05 11:00:01.900676 master-0 kubenswrapper[24928]: I1205 11:00:01.900618 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cn6b" event={"ID":"fec8fc95-cfe1-4f43-b693-7a11680da86a","Type":"ContainerStarted","Data":"ae30539ac9cc6fef7b620921c48dffc9a96581edba2f0f247e4750a730fbeecc"} Dec 05 11:00:01.900676 master-0 kubenswrapper[24928]: I1205 11:00:01.900635 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cn6b" event={"ID":"fec8fc95-cfe1-4f43-b693-7a11680da86a","Type":"ContainerStarted","Data":"9116dc0c2ac62fd43bf729641a851d2c43ab7023ba51602cde3be1c94695ea95"} Dec 05 11:00:01.900676 master-0 kubenswrapper[24928]: I1205 11:00:01.900648 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cn6b" event={"ID":"fec8fc95-cfe1-4f43-b693-7a11680da86a","Type":"ContainerStarted","Data":"7f123ac3b235be4cca7d9160cc948a9e114a17fa12c0ad6a1fcc2144b064c71c"} Dec 05 11:00:01.904065 master-0 kubenswrapper[24928]: I1205 11:00:01.904031 24928 generic.go:334] "Generic (PLEG): container finished" podID="06b902d6-85f1-4661-8cc6-be82f74a966d" containerID="77c9b6fd5d013100eb68e4afeb5ea7a33da6bb0eb6942ac8e68f5e51cd472646" exitCode=0 Dec 05 11:00:01.905170 master-0 kubenswrapper[24928]: I1205 11:00:01.905123 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" event={"ID":"06b902d6-85f1-4661-8cc6-be82f74a966d","Type":"ContainerDied","Data":"77c9b6fd5d013100eb68e4afeb5ea7a33da6bb0eb6942ac8e68f5e51cd472646"} Dec 05 11:00:01.905261 master-0 kubenswrapper[24928]: I1205 11:00:01.905178 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" event={"ID":"06b902d6-85f1-4661-8cc6-be82f74a966d","Type":"ContainerStarted","Data":"bae2332908d8ec4b8eb1b4d370de3e474943572f611427d4b9ecd6322f51be96"} Dec 05 11:00:01.905890 master-0 kubenswrapper[24928]: I1205 11:00:01.905850 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" Dec 05 11:00:01.912292 master-0 kubenswrapper[24928]: I1205 11:00:01.912197 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-hvbdn" podStartSLOduration=7.602549359 podStartE2EDuration="9.912178056s" podCreationTimestamp="2025-12-05 10:59:52 +0000 UTC" firstStartedPulling="2025-12-05 10:59:58.796744561 +0000 UTC m=+758.799938412" lastFinishedPulling="2025-12-05 11:00:01.106373258 +0000 UTC m=+761.109567109" observedRunningTime="2025-12-05 11:00:01.907085302 +0000 UTC m=+761.910279163" watchObservedRunningTime="2025-12-05 11:00:01.912178056 +0000 UTC m=+761.915371917" Dec 05 11:00:02.914806 master-0 kubenswrapper[24928]: I1205 11:00:02.914753 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cn6b" event={"ID":"fec8fc95-cfe1-4f43-b693-7a11680da86a","Type":"ContainerStarted","Data":"3c23182e6681030a7f2586d3db1032940d9e981f2e6dc35ed8f487802ebf6c62"} Dec 05 11:00:02.915917 master-0 kubenswrapper[24928]: I1205 11:00:02.915893 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-2cn6b" Dec 05 11:00:02.916017 master-0 kubenswrapper[24928]: I1205 11:00:02.915998 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-2cn6b" event={"ID":"fec8fc95-cfe1-4f43-b693-7a11680da86a","Type":"ContainerStarted","Data":"728342d777e8e2f2b97795aa257b1ed7c723b1887253702420c59ffaefe5a171"} Dec 05 11:00:02.956608 master-0 kubenswrapper[24928]: I1205 11:00:02.953748 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-2cn6b" podStartSLOduration=5.9446246689999995 podStartE2EDuration="13.953710129s" podCreationTimestamp="2025-12-05 10:59:49 +0000 UTC" firstStartedPulling="2025-12-05 10:59:50.339022012 +0000 UTC m=+750.342215873" lastFinishedPulling="2025-12-05 10:59:58.348107482 +0000 UTC m=+758.351301333" observedRunningTime="2025-12-05 11:00:02.943043688 +0000 UTC m=+762.946237559" watchObservedRunningTime="2025-12-05 11:00:02.953710129 +0000 UTC m=+762.956903990" Dec 05 11:00:03.259240 master-0 kubenswrapper[24928]: I1205 11:00:03.259183 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:03.324960 master-0 kubenswrapper[24928]: I1205 11:00:03.324871 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06b902d6-85f1-4661-8cc6-be82f74a966d-secret-volume\") pod \"06b902d6-85f1-4661-8cc6-be82f74a966d\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " Dec 05 11:00:03.324960 master-0 kubenswrapper[24928]: I1205 11:00:03.324956 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmk4g\" (UniqueName: \"kubernetes.io/projected/06b902d6-85f1-4661-8cc6-be82f74a966d-kube-api-access-gmk4g\") pod \"06b902d6-85f1-4661-8cc6-be82f74a966d\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " Dec 05 11:00:03.325247 master-0 kubenswrapper[24928]: I1205 11:00:03.325118 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06b902d6-85f1-4661-8cc6-be82f74a966d-config-volume\") pod \"06b902d6-85f1-4661-8cc6-be82f74a966d\" (UID: \"06b902d6-85f1-4661-8cc6-be82f74a966d\") " Dec 05 11:00:03.325996 master-0 kubenswrapper[24928]: I1205 11:00:03.325951 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06b902d6-85f1-4661-8cc6-be82f74a966d-config-volume" (OuterVolumeSpecName: "config-volume") pod "06b902d6-85f1-4661-8cc6-be82f74a966d" (UID: "06b902d6-85f1-4661-8cc6-be82f74a966d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:00:03.328368 master-0 kubenswrapper[24928]: I1205 11:00:03.328298 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06b902d6-85f1-4661-8cc6-be82f74a966d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "06b902d6-85f1-4661-8cc6-be82f74a966d" (UID: "06b902d6-85f1-4661-8cc6-be82f74a966d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:00:03.329195 master-0 kubenswrapper[24928]: I1205 11:00:03.329146 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06b902d6-85f1-4661-8cc6-be82f74a966d-kube-api-access-gmk4g" (OuterVolumeSpecName: "kube-api-access-gmk4g") pod "06b902d6-85f1-4661-8cc6-be82f74a966d" (UID: "06b902d6-85f1-4661-8cc6-be82f74a966d"). InnerVolumeSpecName "kube-api-access-gmk4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:00:03.426732 master-0 kubenswrapper[24928]: I1205 11:00:03.426616 24928 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/06b902d6-85f1-4661-8cc6-be82f74a966d-config-volume\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:03.426732 master-0 kubenswrapper[24928]: I1205 11:00:03.426690 24928 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/06b902d6-85f1-4661-8cc6-be82f74a966d-secret-volume\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:03.426732 master-0 kubenswrapper[24928]: I1205 11:00:03.426704 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmk4g\" (UniqueName: \"kubernetes.io/projected/06b902d6-85f1-4661-8cc6-be82f74a966d-kube-api-access-gmk4g\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:03.929161 master-0 kubenswrapper[24928]: I1205 11:00:03.929095 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" Dec 05 11:00:03.929161 master-0 kubenswrapper[24928]: I1205 11:00:03.929080 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm" event={"ID":"06b902d6-85f1-4661-8cc6-be82f74a966d","Type":"ContainerDied","Data":"bae2332908d8ec4b8eb1b4d370de3e474943572f611427d4b9ecd6322f51be96"} Dec 05 11:00:03.929814 master-0 kubenswrapper[24928]: I1205 11:00:03.929184 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bae2332908d8ec4b8eb1b4d370de3e474943572f611427d4b9ecd6322f51be96" Dec 05 11:00:05.037193 master-0 kubenswrapper[24928]: I1205 11:00:05.037113 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-2cn6b" Dec 05 11:00:05.083515 master-0 kubenswrapper[24928]: I1205 11:00:05.083386 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-2cn6b" Dec 05 11:00:05.108799 master-0 kubenswrapper[24928]: I1205 11:00:05.108663 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-78d584df9-x54pl" Dec 05 11:00:05.108799 master-0 kubenswrapper[24928]: I1205 11:00:05.108778 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-78d584df9-x54pl" Dec 05 11:00:05.113489 master-0 kubenswrapper[24928]: I1205 11:00:05.113434 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-78d584df9-x54pl" Dec 05 11:00:05.950970 master-0 kubenswrapper[24928]: I1205 11:00:05.950903 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-78d584df9-x54pl" Dec 05 11:00:06.106128 master-0 kubenswrapper[24928]: I1205 11:00:06.106049 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-d656f4996-kjkt5"] Dec 05 11:00:09.765776 master-0 kubenswrapper[24928]: I1205 11:00:09.765710 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-hxkln" Dec 05 11:00:10.031528 master-0 kubenswrapper[24928]: I1205 11:00:10.031365 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-dlsnb" Dec 05 11:00:14.736612 master-0 kubenswrapper[24928]: I1205 11:00:14.736545 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-jxlrb" Dec 05 11:00:20.043845 master-0 kubenswrapper[24928]: I1205 11:00:20.043704 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-2cn6b" Dec 05 11:00:20.272115 master-0 kubenswrapper[24928]: I1205 11:00:20.272047 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-storage/vg-manager-kf8hp"] Dec 05 11:00:20.272478 master-0 kubenswrapper[24928]: E1205 11:00:20.272450 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06b902d6-85f1-4661-8cc6-be82f74a966d" containerName="collect-profiles" Dec 05 11:00:20.272478 master-0 kubenswrapper[24928]: I1205 11:00:20.272473 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="06b902d6-85f1-4661-8cc6-be82f74a966d" containerName="collect-profiles" Dec 05 11:00:20.272723 master-0 kubenswrapper[24928]: I1205 11:00:20.272696 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="06b902d6-85f1-4661-8cc6-be82f74a966d" containerName="collect-profiles" Dec 05 11:00:20.273331 master-0 kubenswrapper[24928]: I1205 11:00:20.273299 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.275352 master-0 kubenswrapper[24928]: I1205 11:00:20.275285 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-storage"/"vg-manager-metrics-cert" Dec 05 11:00:20.288572 master-0 kubenswrapper[24928]: I1205 11:00:20.288520 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-storage/vg-manager-kf8hp"] Dec 05 11:00:20.390263 master-0 kubenswrapper[24928]: I1205 11:00:20.390100 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9fd6\" (UniqueName: \"kubernetes.io/projected/5254fcbc-d5ca-44e9-89c8-b0856163f22d-kube-api-access-p9fd6\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.390263 master-0 kubenswrapper[24928]: I1205 11:00:20.390192 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-volumes-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-pod-volumes-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.390661 master-0 kubenswrapper[24928]: I1205 11:00:20.390606 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-registration-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.390722 master-0 kubenswrapper[24928]: I1205 11:00:20.390702 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-sys\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.390808 master-0 kubenswrapper[24928]: I1205 11:00:20.390783 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-udev\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-run-udev\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.390880 master-0 kubenswrapper[24928]: I1205 11:00:20.390859 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/5254fcbc-d5ca-44e9-89c8-b0856163f22d-metrics-cert\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.390989 master-0 kubenswrapper[24928]: I1205 11:00:20.390961 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lvmd-config\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-lvmd-config\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.391113 master-0 kubenswrapper[24928]: I1205 11:00:20.391085 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-plugin-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-csi-plugin-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.391174 master-0 kubenswrapper[24928]: I1205 11:00:20.391145 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"file-lock-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-file-lock-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.391212 master-0 kubenswrapper[24928]: I1205 11:00:20.391198 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-plugin-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-node-plugin-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.391246 master-0 kubenswrapper[24928]: I1205 11:00:20.391223 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-device-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492228 master-0 kubenswrapper[24928]: I1205 11:00:20.492169 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-plugin-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-csi-plugin-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492228 master-0 kubenswrapper[24928]: I1205 11:00:20.492221 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"file-lock-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-file-lock-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492540 master-0 kubenswrapper[24928]: I1205 11:00:20.492245 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-plugin-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-node-plugin-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492540 master-0 kubenswrapper[24928]: I1205 11:00:20.492268 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-device-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492540 master-0 kubenswrapper[24928]: I1205 11:00:20.492451 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-plugin-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-csi-plugin-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492686 master-0 kubenswrapper[24928]: I1205 11:00:20.492535 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9fd6\" (UniqueName: \"kubernetes.io/projected/5254fcbc-d5ca-44e9-89c8-b0856163f22d-kube-api-access-p9fd6\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492686 master-0 kubenswrapper[24928]: I1205 11:00:20.492598 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-plugin-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-node-plugin-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492686 master-0 kubenswrapper[24928]: I1205 11:00:20.492623 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"device-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-device-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492686 master-0 kubenswrapper[24928]: I1205 11:00:20.492567 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"file-lock-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-file-lock-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492892 master-0 kubenswrapper[24928]: I1205 11:00:20.492693 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-volumes-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-pod-volumes-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492892 master-0 kubenswrapper[24928]: I1205 11:00:20.492746 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-registration-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492892 master-0 kubenswrapper[24928]: I1205 11:00:20.492782 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-sys\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492892 master-0 kubenswrapper[24928]: I1205 11:00:20.492841 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-volumes-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-pod-volumes-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492892 master-0 kubenswrapper[24928]: I1205 11:00:20.492854 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-udev\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-run-udev\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.492892 master-0 kubenswrapper[24928]: I1205 11:00:20.492889 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-udev\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-run-udev\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.493111 master-0 kubenswrapper[24928]: I1205 11:00:20.492915 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-sys\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.493111 master-0 kubenswrapper[24928]: I1205 11:00:20.492856 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-registration-dir\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.493111 master-0 kubenswrapper[24928]: I1205 11:00:20.492956 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/5254fcbc-d5ca-44e9-89c8-b0856163f22d-metrics-cert\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.493111 master-0 kubenswrapper[24928]: I1205 11:00:20.493036 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lvmd-config\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-lvmd-config\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.493262 master-0 kubenswrapper[24928]: I1205 11:00:20.493235 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lvmd-config\" (UniqueName: \"kubernetes.io/host-path/5254fcbc-d5ca-44e9-89c8-b0856163f22d-lvmd-config\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.496719 master-0 kubenswrapper[24928]: I1205 11:00:20.496690 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-cert\" (UniqueName: \"kubernetes.io/secret/5254fcbc-d5ca-44e9-89c8-b0856163f22d-metrics-cert\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.517315 master-0 kubenswrapper[24928]: I1205 11:00:20.517252 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9fd6\" (UniqueName: \"kubernetes.io/projected/5254fcbc-d5ca-44e9-89c8-b0856163f22d-kube-api-access-p9fd6\") pod \"vg-manager-kf8hp\" (UID: \"5254fcbc-d5ca-44e9-89c8-b0856163f22d\") " pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:20.590483 master-0 kubenswrapper[24928]: I1205 11:00:20.590359 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:21.189338 master-0 kubenswrapper[24928]: I1205 11:00:21.189261 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-storage/vg-manager-kf8hp"] Dec 05 11:00:22.107711 master-0 kubenswrapper[24928]: I1205 11:00:22.107613 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-storage/vg-manager-kf8hp" event={"ID":"5254fcbc-d5ca-44e9-89c8-b0856163f22d","Type":"ContainerStarted","Data":"fc636dd34bbc0e05fe1ff73c33c5a8b0fcafcee12523b73cf283e3a110b0e260"} Dec 05 11:00:22.107711 master-0 kubenswrapper[24928]: I1205 11:00:22.107679 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-storage/vg-manager-kf8hp" event={"ID":"5254fcbc-d5ca-44e9-89c8-b0856163f22d","Type":"ContainerStarted","Data":"e0fd18c179da6a3db572e694dd89c41aa936fa9d720885c92dbba8f4f34fcfe2"} Dec 05 11:00:22.136901 master-0 kubenswrapper[24928]: I1205 11:00:22.136766 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-storage/vg-manager-kf8hp" podStartSLOduration=2.136735165 podStartE2EDuration="2.136735165s" podCreationTimestamp="2025-12-05 11:00:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:00:22.128632666 +0000 UTC m=+782.131826537" watchObservedRunningTime="2025-12-05 11:00:22.136735165 +0000 UTC m=+782.139929066" Dec 05 11:00:24.126729 master-0 kubenswrapper[24928]: I1205 11:00:24.126616 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-storage_vg-manager-kf8hp_5254fcbc-d5ca-44e9-89c8-b0856163f22d/vg-manager/0.log" Dec 05 11:00:24.127402 master-0 kubenswrapper[24928]: I1205 11:00:24.126760 24928 generic.go:334] "Generic (PLEG): container finished" podID="5254fcbc-d5ca-44e9-89c8-b0856163f22d" containerID="fc636dd34bbc0e05fe1ff73c33c5a8b0fcafcee12523b73cf283e3a110b0e260" exitCode=1 Dec 05 11:00:24.127402 master-0 kubenswrapper[24928]: I1205 11:00:24.126831 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-storage/vg-manager-kf8hp" event={"ID":"5254fcbc-d5ca-44e9-89c8-b0856163f22d","Type":"ContainerDied","Data":"fc636dd34bbc0e05fe1ff73c33c5a8b0fcafcee12523b73cf283e3a110b0e260"} Dec 05 11:00:24.128598 master-0 kubenswrapper[24928]: I1205 11:00:24.128552 24928 scope.go:117] "RemoveContainer" containerID="fc636dd34bbc0e05fe1ff73c33c5a8b0fcafcee12523b73cf283e3a110b0e260" Dec 05 11:00:24.513359 master-0 kubenswrapper[24928]: I1205 11:00:24.513218 24928 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/topolvm.io-reg.sock" Dec 05 11:00:25.138206 master-0 kubenswrapper[24928]: I1205 11:00:25.138138 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-storage_vg-manager-kf8hp_5254fcbc-d5ca-44e9-89c8-b0856163f22d/vg-manager/0.log" Dec 05 11:00:25.138730 master-0 kubenswrapper[24928]: I1205 11:00:25.138232 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-storage/vg-manager-kf8hp" event={"ID":"5254fcbc-d5ca-44e9-89c8-b0856163f22d","Type":"ContainerStarted","Data":"e6e5c0c31c54113fc4611f68fe60ae953842b9bea4434589bfe4add794e34f99"} Dec 05 11:00:25.237598 master-0 kubenswrapper[24928]: I1205 11:00:25.237338 24928 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/topolvm.io-reg.sock","Timestamp":"2025-12-05T11:00:24.51327289Z","Handler":null,"Name":""} Dec 05 11:00:25.239616 master-0 kubenswrapper[24928]: I1205 11:00:25.239543 24928 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: topolvm.io endpoint: /var/lib/kubelet/plugins/topolvm.io/node/csi-topolvm.sock versions: 1.0.0 Dec 05 11:00:25.239616 master-0 kubenswrapper[24928]: I1205 11:00:25.239609 24928 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: topolvm.io at endpoint: /var/lib/kubelet/plugins/topolvm.io/node/csi-topolvm.sock Dec 05 11:00:30.593095 master-0 kubenswrapper[24928]: I1205 11:00:30.593032 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:30.598325 master-0 kubenswrapper[24928]: I1205 11:00:30.594037 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:31.164916 master-0 kubenswrapper[24928]: I1205 11:00:31.164765 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-d656f4996-kjkt5" podUID="30aa818a-f5af-4468-9920-fb1f8d1cb811" containerName="console" containerID="cri-o://2ded5fe6d62e26efc3e9628222ed40dc36415ea6ada7ddecc6a0bb5c68225b08" gracePeriod=15 Dec 05 11:00:31.205965 master-0 kubenswrapper[24928]: I1205 11:00:31.205892 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:31.207022 master-0 kubenswrapper[24928]: I1205 11:00:31.206978 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-storage/vg-manager-kf8hp" Dec 05 11:00:32.217124 master-0 kubenswrapper[24928]: I1205 11:00:32.217080 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-d656f4996-kjkt5_30aa818a-f5af-4468-9920-fb1f8d1cb811/console/0.log" Dec 05 11:00:32.217893 master-0 kubenswrapper[24928]: I1205 11:00:32.217863 24928 generic.go:334] "Generic (PLEG): container finished" podID="30aa818a-f5af-4468-9920-fb1f8d1cb811" containerID="2ded5fe6d62e26efc3e9628222ed40dc36415ea6ada7ddecc6a0bb5c68225b08" exitCode=2 Dec 05 11:00:32.218076 master-0 kubenswrapper[24928]: I1205 11:00:32.218003 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-d656f4996-kjkt5" event={"ID":"30aa818a-f5af-4468-9920-fb1f8d1cb811","Type":"ContainerDied","Data":"2ded5fe6d62e26efc3e9628222ed40dc36415ea6ada7ddecc6a0bb5c68225b08"} Dec 05 11:00:33.174450 master-0 kubenswrapper[24928]: I1205 11:00:33.174167 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-d656f4996-kjkt5_30aa818a-f5af-4468-9920-fb1f8d1cb811/console/0.log" Dec 05 11:00:33.174450 master-0 kubenswrapper[24928]: I1205 11:00:33.174263 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-d656f4996-kjkt5" Dec 05 11:00:33.268007 master-0 kubenswrapper[24928]: I1205 11:00:33.267962 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-trusted-ca-bundle\") pod \"30aa818a-f5af-4468-9920-fb1f8d1cb811\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " Dec 05 11:00:33.268776 master-0 kubenswrapper[24928]: I1205 11:00:33.268750 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-serving-cert\") pod \"30aa818a-f5af-4468-9920-fb1f8d1cb811\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " Dec 05 11:00:33.268996 master-0 kubenswrapper[24928]: I1205 11:00:33.268969 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-service-ca\") pod \"30aa818a-f5af-4468-9920-fb1f8d1cb811\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " Dec 05 11:00:33.269184 master-0 kubenswrapper[24928]: I1205 11:00:33.269169 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-oauth-serving-cert\") pod \"30aa818a-f5af-4468-9920-fb1f8d1cb811\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " Dec 05 11:00:33.269311 master-0 kubenswrapper[24928]: I1205 11:00:33.269262 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-oauth-config\") pod \"30aa818a-f5af-4468-9920-fb1f8d1cb811\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " Dec 05 11:00:33.269478 master-0 kubenswrapper[24928]: I1205 11:00:33.269457 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd7s5\" (UniqueName: \"kubernetes.io/projected/30aa818a-f5af-4468-9920-fb1f8d1cb811-kube-api-access-wd7s5\") pod \"30aa818a-f5af-4468-9920-fb1f8d1cb811\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " Dec 05 11:00:33.269671 master-0 kubenswrapper[24928]: I1205 11:00:33.269654 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-config\") pod \"30aa818a-f5af-4468-9920-fb1f8d1cb811\" (UID: \"30aa818a-f5af-4468-9920-fb1f8d1cb811\") " Dec 05 11:00:33.269895 master-0 kubenswrapper[24928]: I1205 11:00:33.269202 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-d656f4996-kjkt5_30aa818a-f5af-4468-9920-fb1f8d1cb811/console/0.log" Dec 05 11:00:33.270024 master-0 kubenswrapper[24928]: I1205 11:00:33.268978 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "30aa818a-f5af-4468-9920-fb1f8d1cb811" (UID: "30aa818a-f5af-4468-9920-fb1f8d1cb811"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:00:33.270024 master-0 kubenswrapper[24928]: I1205 11:00:33.269678 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-service-ca" (OuterVolumeSpecName: "service-ca") pod "30aa818a-f5af-4468-9920-fb1f8d1cb811" (UID: "30aa818a-f5af-4468-9920-fb1f8d1cb811"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:00:33.270024 master-0 kubenswrapper[24928]: I1205 11:00:33.269724 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "30aa818a-f5af-4468-9920-fb1f8d1cb811" (UID: "30aa818a-f5af-4468-9920-fb1f8d1cb811"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:00:33.270180 master-0 kubenswrapper[24928]: I1205 11:00:33.270043 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-d656f4996-kjkt5" event={"ID":"30aa818a-f5af-4468-9920-fb1f8d1cb811","Type":"ContainerDied","Data":"545acad0541c1a9d6068ca98ebe068536bad5bac9c64a9ce82ec7c0d3d590c94"} Dec 05 11:00:33.270180 master-0 kubenswrapper[24928]: I1205 11:00:33.270128 24928 scope.go:117] "RemoveContainer" containerID="2ded5fe6d62e26efc3e9628222ed40dc36415ea6ada7ddecc6a0bb5c68225b08" Dec 05 11:00:33.270644 master-0 kubenswrapper[24928]: I1205 11:00:33.270294 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-d656f4996-kjkt5" Dec 05 11:00:33.270644 master-0 kubenswrapper[24928]: I1205 11:00:33.270485 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-config" (OuterVolumeSpecName: "console-config") pod "30aa818a-f5af-4468-9920-fb1f8d1cb811" (UID: "30aa818a-f5af-4468-9920-fb1f8d1cb811"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:00:33.271215 master-0 kubenswrapper[24928]: I1205 11:00:33.270906 24928 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-oauth-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:33.271215 master-0 kubenswrapper[24928]: I1205 11:00:33.270932 24928 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:33.271215 master-0 kubenswrapper[24928]: I1205 11:00:33.270942 24928 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-trusted-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:33.271215 master-0 kubenswrapper[24928]: I1205 11:00:33.270951 24928 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/30aa818a-f5af-4468-9920-fb1f8d1cb811-service-ca\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:33.274202 master-0 kubenswrapper[24928]: I1205 11:00:33.274051 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "30aa818a-f5af-4468-9920-fb1f8d1cb811" (UID: "30aa818a-f5af-4468-9920-fb1f8d1cb811"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:00:33.274360 master-0 kubenswrapper[24928]: I1205 11:00:33.274243 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30aa818a-f5af-4468-9920-fb1f8d1cb811-kube-api-access-wd7s5" (OuterVolumeSpecName: "kube-api-access-wd7s5") pod "30aa818a-f5af-4468-9920-fb1f8d1cb811" (UID: "30aa818a-f5af-4468-9920-fb1f8d1cb811"). InnerVolumeSpecName "kube-api-access-wd7s5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:00:33.275191 master-0 kubenswrapper[24928]: I1205 11:00:33.275152 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "30aa818a-f5af-4468-9920-fb1f8d1cb811" (UID: "30aa818a-f5af-4468-9920-fb1f8d1cb811"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:00:33.372242 master-0 kubenswrapper[24928]: I1205 11:00:33.372109 24928 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-oauth-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:33.372242 master-0 kubenswrapper[24928]: I1205 11:00:33.372163 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd7s5\" (UniqueName: \"kubernetes.io/projected/30aa818a-f5af-4468-9920-fb1f8d1cb811-kube-api-access-wd7s5\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:33.372242 master-0 kubenswrapper[24928]: I1205 11:00:33.372173 24928 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/30aa818a-f5af-4468-9920-fb1f8d1cb811-console-serving-cert\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:33.624518 master-0 kubenswrapper[24928]: I1205 11:00:33.624039 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-d656f4996-kjkt5"] Dec 05 11:00:33.632173 master-0 kubenswrapper[24928]: I1205 11:00:33.632094 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-d656f4996-kjkt5"] Dec 05 11:00:34.231246 master-0 kubenswrapper[24928]: I1205 11:00:34.231127 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30aa818a-f5af-4468-9920-fb1f8d1cb811" path="/var/lib/kubelet/pods/30aa818a-f5af-4468-9920-fb1f8d1cb811/volumes" Dec 05 11:00:34.982017 master-0 kubenswrapper[24928]: I1205 11:00:34.981932 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-k64dw"] Dec 05 11:00:34.982620 master-0 kubenswrapper[24928]: E1205 11:00:34.982333 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30aa818a-f5af-4468-9920-fb1f8d1cb811" containerName="console" Dec 05 11:00:34.982620 master-0 kubenswrapper[24928]: I1205 11:00:34.982355 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="30aa818a-f5af-4468-9920-fb1f8d1cb811" containerName="console" Dec 05 11:00:34.982620 master-0 kubenswrapper[24928]: I1205 11:00:34.982615 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="30aa818a-f5af-4468-9920-fb1f8d1cb811" containerName="console" Dec 05 11:00:34.983333 master-0 kubenswrapper[24928]: I1205 11:00:34.983291 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k64dw" Dec 05 11:00:34.984806 master-0 kubenswrapper[24928]: I1205 11:00:34.984744 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 05 11:00:34.985053 master-0 kubenswrapper[24928]: I1205 11:00:34.985025 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 05 11:00:34.996008 master-0 kubenswrapper[24928]: I1205 11:00:34.995944 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k64dw"] Dec 05 11:00:35.101503 master-0 kubenswrapper[24928]: I1205 11:00:35.101415 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97m94\" (UniqueName: \"kubernetes.io/projected/6b418fb9-a9c4-46b7-9be4-0dc52e339e87-kube-api-access-97m94\") pod \"openstack-operator-index-k64dw\" (UID: \"6b418fb9-a9c4-46b7-9be4-0dc52e339e87\") " pod="openstack-operators/openstack-operator-index-k64dw" Dec 05 11:00:35.203058 master-0 kubenswrapper[24928]: I1205 11:00:35.202982 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97m94\" (UniqueName: \"kubernetes.io/projected/6b418fb9-a9c4-46b7-9be4-0dc52e339e87-kube-api-access-97m94\") pod \"openstack-operator-index-k64dw\" (UID: \"6b418fb9-a9c4-46b7-9be4-0dc52e339e87\") " pod="openstack-operators/openstack-operator-index-k64dw" Dec 05 11:00:35.222749 master-0 kubenswrapper[24928]: I1205 11:00:35.222689 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97m94\" (UniqueName: \"kubernetes.io/projected/6b418fb9-a9c4-46b7-9be4-0dc52e339e87-kube-api-access-97m94\") pod \"openstack-operator-index-k64dw\" (UID: \"6b418fb9-a9c4-46b7-9be4-0dc52e339e87\") " pod="openstack-operators/openstack-operator-index-k64dw" Dec 05 11:00:35.358804 master-0 kubenswrapper[24928]: I1205 11:00:35.358727 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-k64dw" Dec 05 11:00:35.807519 master-0 kubenswrapper[24928]: I1205 11:00:35.807466 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-k64dw"] Dec 05 11:00:36.298710 master-0 kubenswrapper[24928]: I1205 11:00:36.298586 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k64dw" event={"ID":"6b418fb9-a9c4-46b7-9be4-0dc52e339e87","Type":"ContainerStarted","Data":"64ed4bd263189c41584f529a0ce02167ef79bb673e53f5e9183a0e65c0eb1644"} Dec 05 11:00:37.309092 master-0 kubenswrapper[24928]: I1205 11:00:37.308973 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-k64dw" event={"ID":"6b418fb9-a9c4-46b7-9be4-0dc52e339e87","Type":"ContainerStarted","Data":"1023a66d3a1812f2a83fb582e35950b060d5cafa51ca3e91742c6ae850376f69"} Dec 05 11:00:37.335497 master-0 kubenswrapper[24928]: I1205 11:00:37.335361 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-k64dw" podStartSLOduration=2.553467915 podStartE2EDuration="3.335340267s" podCreationTimestamp="2025-12-05 11:00:34 +0000 UTC" firstStartedPulling="2025-12-05 11:00:35.81602451 +0000 UTC m=+795.819218381" lastFinishedPulling="2025-12-05 11:00:36.597896882 +0000 UTC m=+796.601090733" observedRunningTime="2025-12-05 11:00:37.335270125 +0000 UTC m=+797.338464056" watchObservedRunningTime="2025-12-05 11:00:37.335340267 +0000 UTC m=+797.338534118" Dec 05 11:00:45.359361 master-0 kubenswrapper[24928]: I1205 11:00:45.359222 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-k64dw" Dec 05 11:00:45.359361 master-0 kubenswrapper[24928]: I1205 11:00:45.359330 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-k64dw" Dec 05 11:00:45.417334 master-0 kubenswrapper[24928]: I1205 11:00:45.417262 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-k64dw" Dec 05 11:00:45.453294 master-0 kubenswrapper[24928]: I1205 11:00:45.453213 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-k64dw" Dec 05 11:00:51.564749 master-0 kubenswrapper[24928]: I1205 11:00:51.564678 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7"] Dec 05 11:00:51.566408 master-0 kubenswrapper[24928]: I1205 11:00:51.566376 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:51.583801 master-0 kubenswrapper[24928]: I1205 11:00:51.583725 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7"] Dec 05 11:00:51.711915 master-0 kubenswrapper[24928]: I1205 11:00:51.711810 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:51.712156 master-0 kubenswrapper[24928]: I1205 11:00:51.711985 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbcqq\" (UniqueName: \"kubernetes.io/projected/0fb13d51-751c-4313-99f3-27db02ea0911-kube-api-access-jbcqq\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:51.712156 master-0 kubenswrapper[24928]: I1205 11:00:51.712029 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:51.813487 master-0 kubenswrapper[24928]: I1205 11:00:51.813387 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbcqq\" (UniqueName: \"kubernetes.io/projected/0fb13d51-751c-4313-99f3-27db02ea0911-kube-api-access-jbcqq\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:51.813739 master-0 kubenswrapper[24928]: I1205 11:00:51.813679 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:51.814078 master-0 kubenswrapper[24928]: I1205 11:00:51.814031 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:51.814234 master-0 kubenswrapper[24928]: I1205 11:00:51.814193 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:51.814525 master-0 kubenswrapper[24928]: I1205 11:00:51.814475 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:51.835886 master-0 kubenswrapper[24928]: I1205 11:00:51.835711 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbcqq\" (UniqueName: \"kubernetes.io/projected/0fb13d51-751c-4313-99f3-27db02ea0911-kube-api-access-jbcqq\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:51.884204 master-0 kubenswrapper[24928]: I1205 11:00:51.884099 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:53.289191 master-0 kubenswrapper[24928]: I1205 11:00:53.289108 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7"] Dec 05 11:00:53.290779 master-0 kubenswrapper[24928]: W1205 11:00:53.290731 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fb13d51_751c_4313_99f3_27db02ea0911.slice/crio-6ce0ef457c25f67b1ff473028fbc4ee6ad6359474ceb998b4bc6cef2129f23f2 WatchSource:0}: Error finding container 6ce0ef457c25f67b1ff473028fbc4ee6ad6359474ceb998b4bc6cef2129f23f2: Status 404 returned error can't find the container with id 6ce0ef457c25f67b1ff473028fbc4ee6ad6359474ceb998b4bc6cef2129f23f2 Dec 05 11:00:53.454956 master-0 kubenswrapper[24928]: I1205 11:00:53.454880 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" event={"ID":"0fb13d51-751c-4313-99f3-27db02ea0911","Type":"ContainerStarted","Data":"6ce0ef457c25f67b1ff473028fbc4ee6ad6359474ceb998b4bc6cef2129f23f2"} Dec 05 11:00:54.464141 master-0 kubenswrapper[24928]: I1205 11:00:54.464066 24928 generic.go:334] "Generic (PLEG): container finished" podID="0fb13d51-751c-4313-99f3-27db02ea0911" containerID="9258b13a2e3dcdefa92f772ed833d2cab638ad6eb5306cfaa4f02a9a0ffe547b" exitCode=0 Dec 05 11:00:54.465185 master-0 kubenswrapper[24928]: I1205 11:00:54.464139 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" event={"ID":"0fb13d51-751c-4313-99f3-27db02ea0911","Type":"ContainerDied","Data":"9258b13a2e3dcdefa92f772ed833d2cab638ad6eb5306cfaa4f02a9a0ffe547b"} Dec 05 11:00:56.488171 master-0 kubenswrapper[24928]: I1205 11:00:56.485977 24928 generic.go:334] "Generic (PLEG): container finished" podID="0fb13d51-751c-4313-99f3-27db02ea0911" containerID="9a90c2171408d8fd01ab83a98f586619c0ee5e9ac79ddb324d83f0ccc9f9cb88" exitCode=0 Dec 05 11:00:56.488171 master-0 kubenswrapper[24928]: I1205 11:00:56.486049 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" event={"ID":"0fb13d51-751c-4313-99f3-27db02ea0911","Type":"ContainerDied","Data":"9a90c2171408d8fd01ab83a98f586619c0ee5e9ac79ddb324d83f0ccc9f9cb88"} Dec 05 11:00:57.495553 master-0 kubenswrapper[24928]: I1205 11:00:57.495493 24928 generic.go:334] "Generic (PLEG): container finished" podID="0fb13d51-751c-4313-99f3-27db02ea0911" containerID="0b432e452a35fa5e6635644458977feeb78e3c5f990dca04752936960bf46a3c" exitCode=0 Dec 05 11:00:57.496389 master-0 kubenswrapper[24928]: I1205 11:00:57.495552 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" event={"ID":"0fb13d51-751c-4313-99f3-27db02ea0911","Type":"ContainerDied","Data":"0b432e452a35fa5e6635644458977feeb78e3c5f990dca04752936960bf46a3c"} Dec 05 11:00:58.887270 master-0 kubenswrapper[24928]: I1205 11:00:58.887196 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:00:59.085464 master-0 kubenswrapper[24928]: I1205 11:00:59.085184 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-util\") pod \"0fb13d51-751c-4313-99f3-27db02ea0911\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " Dec 05 11:00:59.085464 master-0 kubenswrapper[24928]: I1205 11:00:59.085294 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbcqq\" (UniqueName: \"kubernetes.io/projected/0fb13d51-751c-4313-99f3-27db02ea0911-kube-api-access-jbcqq\") pod \"0fb13d51-751c-4313-99f3-27db02ea0911\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " Dec 05 11:00:59.085464 master-0 kubenswrapper[24928]: I1205 11:00:59.085343 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-bundle\") pod \"0fb13d51-751c-4313-99f3-27db02ea0911\" (UID: \"0fb13d51-751c-4313-99f3-27db02ea0911\") " Dec 05 11:00:59.087566 master-0 kubenswrapper[24928]: I1205 11:00:59.086765 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-bundle" (OuterVolumeSpecName: "bundle") pod "0fb13d51-751c-4313-99f3-27db02ea0911" (UID: "0fb13d51-751c-4313-99f3-27db02ea0911"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:00:59.092613 master-0 kubenswrapper[24928]: I1205 11:00:59.092516 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fb13d51-751c-4313-99f3-27db02ea0911-kube-api-access-jbcqq" (OuterVolumeSpecName: "kube-api-access-jbcqq") pod "0fb13d51-751c-4313-99f3-27db02ea0911" (UID: "0fb13d51-751c-4313-99f3-27db02ea0911"). InnerVolumeSpecName "kube-api-access-jbcqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:00:59.098407 master-0 kubenswrapper[24928]: I1205 11:00:59.098246 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-util" (OuterVolumeSpecName: "util") pod "0fb13d51-751c-4313-99f3-27db02ea0911" (UID: "0fb13d51-751c-4313-99f3-27db02ea0911"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:00:59.189062 master-0 kubenswrapper[24928]: I1205 11:00:59.188924 24928 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-util\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:59.189062 master-0 kubenswrapper[24928]: I1205 11:00:59.189039 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbcqq\" (UniqueName: \"kubernetes.io/projected/0fb13d51-751c-4313-99f3-27db02ea0911-kube-api-access-jbcqq\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:59.189062 master-0 kubenswrapper[24928]: I1205 11:00:59.189060 24928 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0fb13d51-751c-4313-99f3-27db02ea0911-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:00:59.514155 master-0 kubenswrapper[24928]: I1205 11:00:59.513998 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" event={"ID":"0fb13d51-751c-4313-99f3-27db02ea0911","Type":"ContainerDied","Data":"6ce0ef457c25f67b1ff473028fbc4ee6ad6359474ceb998b4bc6cef2129f23f2"} Dec 05 11:00:59.514155 master-0 kubenswrapper[24928]: I1205 11:00:59.514062 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ce0ef457c25f67b1ff473028fbc4ee6ad6359474ceb998b4bc6cef2129f23f2" Dec 05 11:00:59.514155 master-0 kubenswrapper[24928]: I1205 11:00:59.514087 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7" Dec 05 11:01:06.300039 master-0 kubenswrapper[24928]: I1205 11:01:06.299953 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv"] Dec 05 11:01:06.300716 master-0 kubenswrapper[24928]: E1205 11:01:06.300385 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb13d51-751c-4313-99f3-27db02ea0911" containerName="util" Dec 05 11:01:06.300716 master-0 kubenswrapper[24928]: I1205 11:01:06.300407 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb13d51-751c-4313-99f3-27db02ea0911" containerName="util" Dec 05 11:01:06.300716 master-0 kubenswrapper[24928]: E1205 11:01:06.300453 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb13d51-751c-4313-99f3-27db02ea0911" containerName="extract" Dec 05 11:01:06.300716 master-0 kubenswrapper[24928]: I1205 11:01:06.300465 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb13d51-751c-4313-99f3-27db02ea0911" containerName="extract" Dec 05 11:01:06.300716 master-0 kubenswrapper[24928]: E1205 11:01:06.300506 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb13d51-751c-4313-99f3-27db02ea0911" containerName="pull" Dec 05 11:01:06.300716 master-0 kubenswrapper[24928]: I1205 11:01:06.300515 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb13d51-751c-4313-99f3-27db02ea0911" containerName="pull" Dec 05 11:01:06.300716 master-0 kubenswrapper[24928]: I1205 11:01:06.300690 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fb13d51-751c-4313-99f3-27db02ea0911" containerName="extract" Dec 05 11:01:06.303205 master-0 kubenswrapper[24928]: I1205 11:01:06.303008 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" Dec 05 11:01:06.326113 master-0 kubenswrapper[24928]: I1205 11:01:06.324187 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c59cw\" (UniqueName: \"kubernetes.io/projected/db78402f-ed15-45d6-8f62-18e4d684efbc-kube-api-access-c59cw\") pod \"openstack-operator-controller-operator-55b6fb9447-lq5vv\" (UID: \"db78402f-ed15-45d6-8f62-18e4d684efbc\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" Dec 05 11:01:06.326113 master-0 kubenswrapper[24928]: I1205 11:01:06.324926 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv"] Dec 05 11:01:06.427795 master-0 kubenswrapper[24928]: I1205 11:01:06.427644 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c59cw\" (UniqueName: \"kubernetes.io/projected/db78402f-ed15-45d6-8f62-18e4d684efbc-kube-api-access-c59cw\") pod \"openstack-operator-controller-operator-55b6fb9447-lq5vv\" (UID: \"db78402f-ed15-45d6-8f62-18e4d684efbc\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" Dec 05 11:01:06.451351 master-0 kubenswrapper[24928]: I1205 11:01:06.451288 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c59cw\" (UniqueName: \"kubernetes.io/projected/db78402f-ed15-45d6-8f62-18e4d684efbc-kube-api-access-c59cw\") pod \"openstack-operator-controller-operator-55b6fb9447-lq5vv\" (UID: \"db78402f-ed15-45d6-8f62-18e4d684efbc\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" Dec 05 11:01:06.641211 master-0 kubenswrapper[24928]: I1205 11:01:06.641015 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" Dec 05 11:01:07.089604 master-0 kubenswrapper[24928]: I1205 11:01:07.089552 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv"] Dec 05 11:01:07.588356 master-0 kubenswrapper[24928]: I1205 11:01:07.588270 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" event={"ID":"db78402f-ed15-45d6-8f62-18e4d684efbc","Type":"ContainerStarted","Data":"9bf9cac5b01dc7af85d3225a046e876688b59738e0a48062fd495057f97a4d1e"} Dec 05 11:01:11.654886 master-0 kubenswrapper[24928]: I1205 11:01:11.654733 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" event={"ID":"db78402f-ed15-45d6-8f62-18e4d684efbc","Type":"ContainerStarted","Data":"2255479d7d770385c17278b35b775194ce313ffe6a7b7e41989d66922fc30057"} Dec 05 11:01:11.655495 master-0 kubenswrapper[24928]: I1205 11:01:11.654916 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" Dec 05 11:01:11.691596 master-0 kubenswrapper[24928]: I1205 11:01:11.691524 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" podStartSLOduration=1.395339611 podStartE2EDuration="5.69150501s" podCreationTimestamp="2025-12-05 11:01:06 +0000 UTC" firstStartedPulling="2025-12-05 11:01:07.087596372 +0000 UTC m=+827.090790223" lastFinishedPulling="2025-12-05 11:01:11.383761771 +0000 UTC m=+831.386955622" observedRunningTime="2025-12-05 11:01:11.6899147 +0000 UTC m=+831.693108571" watchObservedRunningTime="2025-12-05 11:01:11.69150501 +0000 UTC m=+831.694698861" Dec 05 11:01:16.644631 master-0 kubenswrapper[24928]: I1205 11:01:16.644546 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" Dec 05 11:01:20.146079 master-0 kubenswrapper[24928]: I1205 11:01:20.146016 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s"] Dec 05 11:01:20.147401 master-0 kubenswrapper[24928]: I1205 11:01:20.147372 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s" Dec 05 11:01:20.175371 master-0 kubenswrapper[24928]: I1205 11:01:20.175323 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s"] Dec 05 11:01:20.198646 master-0 kubenswrapper[24928]: I1205 11:01:20.198579 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gstpt\" (UniqueName: \"kubernetes.io/projected/6876062e-bc77-4979-8057-f6b5f389fdcd-kube-api-access-gstpt\") pod \"openstack-operator-controller-operator-589d7b4556-v294s\" (UID: \"6876062e-bc77-4979-8057-f6b5f389fdcd\") " pod="openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s" Dec 05 11:01:20.301310 master-0 kubenswrapper[24928]: I1205 11:01:20.301227 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gstpt\" (UniqueName: \"kubernetes.io/projected/6876062e-bc77-4979-8057-f6b5f389fdcd-kube-api-access-gstpt\") pod \"openstack-operator-controller-operator-589d7b4556-v294s\" (UID: \"6876062e-bc77-4979-8057-f6b5f389fdcd\") " pod="openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s" Dec 05 11:01:20.323138 master-0 kubenswrapper[24928]: I1205 11:01:20.323094 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gstpt\" (UniqueName: \"kubernetes.io/projected/6876062e-bc77-4979-8057-f6b5f389fdcd-kube-api-access-gstpt\") pod \"openstack-operator-controller-operator-589d7b4556-v294s\" (UID: \"6876062e-bc77-4979-8057-f6b5f389fdcd\") " pod="openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s" Dec 05 11:01:20.464553 master-0 kubenswrapper[24928]: I1205 11:01:20.464413 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s" Dec 05 11:01:21.004105 master-0 kubenswrapper[24928]: I1205 11:01:21.004020 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s"] Dec 05 11:01:21.755219 master-0 kubenswrapper[24928]: I1205 11:01:21.755137 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s" event={"ID":"6876062e-bc77-4979-8057-f6b5f389fdcd","Type":"ContainerStarted","Data":"e1ee9e8870afbcc5f119a774de31990e51eef9fa2170b80abac166d827590285"} Dec 05 11:01:21.755219 master-0 kubenswrapper[24928]: I1205 11:01:21.755194 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s" event={"ID":"6876062e-bc77-4979-8057-f6b5f389fdcd","Type":"ContainerStarted","Data":"918b622c37db09dec0cf823bc64133f074e431d6750f56a3d00becb942255f19"} Dec 05 11:01:21.756074 master-0 kubenswrapper[24928]: I1205 11:01:21.755324 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s" Dec 05 11:01:21.798754 master-0 kubenswrapper[24928]: I1205 11:01:21.798676 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s" podStartSLOduration=1.7986183580000001 podStartE2EDuration="1.798618358s" podCreationTimestamp="2025-12-05 11:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:01:21.797816429 +0000 UTC m=+841.801010310" watchObservedRunningTime="2025-12-05 11:01:21.798618358 +0000 UTC m=+841.801812209" Dec 05 11:01:30.469189 master-0 kubenswrapper[24928]: I1205 11:01:30.469092 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-589d7b4556-v294s" Dec 05 11:01:30.630748 master-0 kubenswrapper[24928]: I1205 11:01:30.619510 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv"] Dec 05 11:01:30.630748 master-0 kubenswrapper[24928]: I1205 11:01:30.620523 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" podUID="db78402f-ed15-45d6-8f62-18e4d684efbc" containerName="operator" containerID="cri-o://2255479d7d770385c17278b35b775194ce313ffe6a7b7e41989d66922fc30057" gracePeriod=10 Dec 05 11:01:30.881446 master-0 kubenswrapper[24928]: I1205 11:01:30.867535 24928 generic.go:334] "Generic (PLEG): container finished" podID="db78402f-ed15-45d6-8f62-18e4d684efbc" containerID="2255479d7d770385c17278b35b775194ce313ffe6a7b7e41989d66922fc30057" exitCode=0 Dec 05 11:01:30.881446 master-0 kubenswrapper[24928]: I1205 11:01:30.867587 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" event={"ID":"db78402f-ed15-45d6-8f62-18e4d684efbc","Type":"ContainerDied","Data":"2255479d7d770385c17278b35b775194ce313ffe6a7b7e41989d66922fc30057"} Dec 05 11:01:31.123904 master-0 kubenswrapper[24928]: I1205 11:01:31.123831 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" Dec 05 11:01:31.206481 master-0 kubenswrapper[24928]: I1205 11:01:31.206377 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c59cw\" (UniqueName: \"kubernetes.io/projected/db78402f-ed15-45d6-8f62-18e4d684efbc-kube-api-access-c59cw\") pod \"db78402f-ed15-45d6-8f62-18e4d684efbc\" (UID: \"db78402f-ed15-45d6-8f62-18e4d684efbc\") " Dec 05 11:01:31.209964 master-0 kubenswrapper[24928]: I1205 11:01:31.209909 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db78402f-ed15-45d6-8f62-18e4d684efbc-kube-api-access-c59cw" (OuterVolumeSpecName: "kube-api-access-c59cw") pod "db78402f-ed15-45d6-8f62-18e4d684efbc" (UID: "db78402f-ed15-45d6-8f62-18e4d684efbc"). InnerVolumeSpecName "kube-api-access-c59cw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:01:31.308661 master-0 kubenswrapper[24928]: I1205 11:01:31.308597 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c59cw\" (UniqueName: \"kubernetes.io/projected/db78402f-ed15-45d6-8f62-18e4d684efbc-kube-api-access-c59cw\") on node \"master-0\" DevicePath \"\"" Dec 05 11:01:31.876979 master-0 kubenswrapper[24928]: I1205 11:01:31.876920 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" event={"ID":"db78402f-ed15-45d6-8f62-18e4d684efbc","Type":"ContainerDied","Data":"9bf9cac5b01dc7af85d3225a046e876688b59738e0a48062fd495057f97a4d1e"} Dec 05 11:01:31.876979 master-0 kubenswrapper[24928]: I1205 11:01:31.876977 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv" Dec 05 11:01:31.877594 master-0 kubenswrapper[24928]: I1205 11:01:31.876990 24928 scope.go:117] "RemoveContainer" containerID="2255479d7d770385c17278b35b775194ce313ffe6a7b7e41989d66922fc30057" Dec 05 11:01:31.924924 master-0 kubenswrapper[24928]: I1205 11:01:31.924838 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv"] Dec 05 11:01:31.932031 master-0 kubenswrapper[24928]: I1205 11:01:31.931969 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-lq5vv"] Dec 05 11:01:32.214281 master-0 kubenswrapper[24928]: I1205 11:01:32.214147 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db78402f-ed15-45d6-8f62-18e4d684efbc" path="/var/lib/kubelet/pods/db78402f-ed15-45d6-8f62-18e4d684efbc/volumes" Dec 05 11:02:30.296318 master-0 kubenswrapper[24928]: I1205 11:02:30.296215 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2"] Dec 05 11:02:30.297078 master-0 kubenswrapper[24928]: E1205 11:02:30.296768 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db78402f-ed15-45d6-8f62-18e4d684efbc" containerName="operator" Dec 05 11:02:30.297078 master-0 kubenswrapper[24928]: I1205 11:02:30.296785 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="db78402f-ed15-45d6-8f62-18e4d684efbc" containerName="operator" Dec 05 11:02:30.297078 master-0 kubenswrapper[24928]: I1205 11:02:30.297011 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="db78402f-ed15-45d6-8f62-18e4d684efbc" containerName="operator" Dec 05 11:02:30.298093 master-0 kubenswrapper[24928]: I1205 11:02:30.298057 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2" Dec 05 11:02:30.344982 master-0 kubenswrapper[24928]: I1205 11:02:30.321810 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v"] Dec 05 11:02:30.344982 master-0 kubenswrapper[24928]: I1205 11:02:30.323466 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v" Dec 05 11:02:30.344982 master-0 kubenswrapper[24928]: I1205 11:02:30.332569 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2"] Dec 05 11:02:30.357323 master-0 kubenswrapper[24928]: I1205 11:02:30.357217 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v"] Dec 05 11:02:30.379109 master-0 kubenswrapper[24928]: I1205 11:02:30.379040 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w"] Dec 05 11:02:30.380697 master-0 kubenswrapper[24928]: I1205 11:02:30.380583 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w" Dec 05 11:02:30.395705 master-0 kubenswrapper[24928]: I1205 11:02:30.395651 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w"] Dec 05 11:02:30.406988 master-0 kubenswrapper[24928]: I1205 11:02:30.406924 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zff95\" (UniqueName: \"kubernetes.io/projected/5017646a-fb39-440c-afe5-a4d6cee68ed4-kube-api-access-zff95\") pod \"barbican-operator-controller-manager-5cd89994b5-ssmd2\" (UID: \"5017646a-fb39-440c-afe5-a4d6cee68ed4\") " pod="openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2" Dec 05 11:02:30.407190 master-0 kubenswrapper[24928]: I1205 11:02:30.407125 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szhgs\" (UniqueName: \"kubernetes.io/projected/4f712f48-af26-4222-a231-ac989f78d18f-kube-api-access-szhgs\") pod \"cinder-operator-controller-manager-f8856dd79-7582v\" (UID: \"4f712f48-af26-4222-a231-ac989f78d18f\") " pod="openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v" Dec 05 11:02:30.415959 master-0 kubenswrapper[24928]: I1205 11:02:30.415860 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc"] Dec 05 11:02:30.417932 master-0 kubenswrapper[24928]: I1205 11:02:30.417807 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc" Dec 05 11:02:30.427762 master-0 kubenswrapper[24928]: I1205 11:02:30.427655 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc"] Dec 05 11:02:30.435169 master-0 kubenswrapper[24928]: I1205 11:02:30.434939 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc"] Dec 05 11:02:30.440409 master-0 kubenswrapper[24928]: I1205 11:02:30.440226 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc" Dec 05 11:02:30.530669 master-0 kubenswrapper[24928]: I1205 11:02:30.507666 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc"] Dec 05 11:02:30.530669 master-0 kubenswrapper[24928]: I1205 11:02:30.509333 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zff95\" (UniqueName: \"kubernetes.io/projected/5017646a-fb39-440c-afe5-a4d6cee68ed4-kube-api-access-zff95\") pod \"barbican-operator-controller-manager-5cd89994b5-ssmd2\" (UID: \"5017646a-fb39-440c-afe5-a4d6cee68ed4\") " pod="openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2" Dec 05 11:02:30.530669 master-0 kubenswrapper[24928]: I1205 11:02:30.509400 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lcbw\" (UniqueName: \"kubernetes.io/projected/1d175522-0c08-40c5-a309-f1820406ea2f-kube-api-access-9lcbw\") pod \"heat-operator-controller-manager-7fd96594c7-5k6gc\" (UID: \"1d175522-0c08-40c5-a309-f1820406ea2f\") " pod="openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc" Dec 05 11:02:30.530669 master-0 kubenswrapper[24928]: I1205 11:02:30.509499 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wfcf\" (UniqueName: \"kubernetes.io/projected/44343d25-46da-40a9-b0be-94861421d403-kube-api-access-7wfcf\") pod \"designate-operator-controller-manager-84bc9f68f5-t8l7w\" (UID: \"44343d25-46da-40a9-b0be-94861421d403\") " pod="openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w" Dec 05 11:02:30.530669 master-0 kubenswrapper[24928]: I1205 11:02:30.509533 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pswtt\" (UniqueName: \"kubernetes.io/projected/216bdcf3-b763-4293-8ac4-3e7eeeae0b5d-kube-api-access-pswtt\") pod \"glance-operator-controller-manager-78cd4f7769-xpcsc\" (UID: \"216bdcf3-b763-4293-8ac4-3e7eeeae0b5d\") " pod="openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc" Dec 05 11:02:30.530669 master-0 kubenswrapper[24928]: I1205 11:02:30.509583 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szhgs\" (UniqueName: \"kubernetes.io/projected/4f712f48-af26-4222-a231-ac989f78d18f-kube-api-access-szhgs\") pod \"cinder-operator-controller-manager-f8856dd79-7582v\" (UID: \"4f712f48-af26-4222-a231-ac989f78d18f\") " pod="openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v" Dec 05 11:02:30.543698 master-0 kubenswrapper[24928]: I1205 11:02:30.543647 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szhgs\" (UniqueName: \"kubernetes.io/projected/4f712f48-af26-4222-a231-ac989f78d18f-kube-api-access-szhgs\") pod \"cinder-operator-controller-manager-f8856dd79-7582v\" (UID: \"4f712f48-af26-4222-a231-ac989f78d18f\") " pod="openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v" Dec 05 11:02:30.552710 master-0 kubenswrapper[24928]: I1205 11:02:30.550819 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zff95\" (UniqueName: \"kubernetes.io/projected/5017646a-fb39-440c-afe5-a4d6cee68ed4-kube-api-access-zff95\") pod \"barbican-operator-controller-manager-5cd89994b5-ssmd2\" (UID: \"5017646a-fb39-440c-afe5-a4d6cee68ed4\") " pod="openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2" Dec 05 11:02:30.564096 master-0 kubenswrapper[24928]: I1205 11:02:30.556163 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c"] Dec 05 11:02:30.564096 master-0 kubenswrapper[24928]: I1205 11:02:30.559117 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c" Dec 05 11:02:30.599108 master-0 kubenswrapper[24928]: I1205 11:02:30.598555 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c"] Dec 05 11:02:30.612084 master-0 kubenswrapper[24928]: I1205 11:02:30.611286 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wfcf\" (UniqueName: \"kubernetes.io/projected/44343d25-46da-40a9-b0be-94861421d403-kube-api-access-7wfcf\") pod \"designate-operator-controller-manager-84bc9f68f5-t8l7w\" (UID: \"44343d25-46da-40a9-b0be-94861421d403\") " pod="openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w" Dec 05 11:02:30.612084 master-0 kubenswrapper[24928]: I1205 11:02:30.611341 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pswtt\" (UniqueName: \"kubernetes.io/projected/216bdcf3-b763-4293-8ac4-3e7eeeae0b5d-kube-api-access-pswtt\") pod \"glance-operator-controller-manager-78cd4f7769-xpcsc\" (UID: \"216bdcf3-b763-4293-8ac4-3e7eeeae0b5d\") " pod="openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc" Dec 05 11:02:30.612084 master-0 kubenswrapper[24928]: I1205 11:02:30.611440 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdrqs\" (UniqueName: \"kubernetes.io/projected/be90920f-159a-4fac-9746-70846d306f22-kube-api-access-pdrqs\") pod \"horizon-operator-controller-manager-f6cc97788-5lr6c\" (UID: \"be90920f-159a-4fac-9746-70846d306f22\") " pod="openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c" Dec 05 11:02:30.612084 master-0 kubenswrapper[24928]: I1205 11:02:30.611468 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lcbw\" (UniqueName: \"kubernetes.io/projected/1d175522-0c08-40c5-a309-f1820406ea2f-kube-api-access-9lcbw\") pod \"heat-operator-controller-manager-7fd96594c7-5k6gc\" (UID: \"1d175522-0c08-40c5-a309-f1820406ea2f\") " pod="openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc" Dec 05 11:02:30.635375 master-0 kubenswrapper[24928]: I1205 11:02:30.635305 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g"] Dec 05 11:02:30.636692 master-0 kubenswrapper[24928]: I1205 11:02:30.636656 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lcbw\" (UniqueName: \"kubernetes.io/projected/1d175522-0c08-40c5-a309-f1820406ea2f-kube-api-access-9lcbw\") pod \"heat-operator-controller-manager-7fd96594c7-5k6gc\" (UID: \"1d175522-0c08-40c5-a309-f1820406ea2f\") " pod="openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc" Dec 05 11:02:30.636816 master-0 kubenswrapper[24928]: I1205 11:02:30.636761 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pswtt\" (UniqueName: \"kubernetes.io/projected/216bdcf3-b763-4293-8ac4-3e7eeeae0b5d-kube-api-access-pswtt\") pod \"glance-operator-controller-manager-78cd4f7769-xpcsc\" (UID: \"216bdcf3-b763-4293-8ac4-3e7eeeae0b5d\") " pod="openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc" Dec 05 11:02:30.637064 master-0 kubenswrapper[24928]: I1205 11:02:30.637017 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:02:30.642185 master-0 kubenswrapper[24928]: I1205 11:02:30.639149 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wfcf\" (UniqueName: \"kubernetes.io/projected/44343d25-46da-40a9-b0be-94861421d403-kube-api-access-7wfcf\") pod \"designate-operator-controller-manager-84bc9f68f5-t8l7w\" (UID: \"44343d25-46da-40a9-b0be-94861421d403\") " pod="openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w" Dec 05 11:02:30.662552 master-0 kubenswrapper[24928]: I1205 11:02:30.642589 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 05 11:02:30.680076 master-0 kubenswrapper[24928]: I1205 11:02:30.677777 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2" Dec 05 11:02:30.699337 master-0 kubenswrapper[24928]: I1205 11:02:30.699274 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v" Dec 05 11:02:30.706777 master-0 kubenswrapper[24928]: I1205 11:02:30.706733 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g"] Dec 05 11:02:30.718188 master-0 kubenswrapper[24928]: I1205 11:02:30.716387 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:02:30.718188 master-0 kubenswrapper[24928]: I1205 11:02:30.716502 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brx8s\" (UniqueName: \"kubernetes.io/projected/da186c71-5e15-45a3-ad76-e51aad43439f-kube-api-access-brx8s\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:02:30.718188 master-0 kubenswrapper[24928]: I1205 11:02:30.716569 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdrqs\" (UniqueName: \"kubernetes.io/projected/be90920f-159a-4fac-9746-70846d306f22-kube-api-access-pdrqs\") pod \"horizon-operator-controller-manager-f6cc97788-5lr6c\" (UID: \"be90920f-159a-4fac-9746-70846d306f22\") " pod="openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c" Dec 05 11:02:30.729902 master-0 kubenswrapper[24928]: I1205 11:02:30.729752 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z"] Dec 05 11:02:30.731687 master-0 kubenswrapper[24928]: I1205 11:02:30.731649 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z" Dec 05 11:02:30.734712 master-0 kubenswrapper[24928]: I1205 11:02:30.734554 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w" Dec 05 11:02:30.750596 master-0 kubenswrapper[24928]: I1205 11:02:30.749173 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4"] Dec 05 11:02:30.750804 master-0 kubenswrapper[24928]: I1205 11:02:30.750763 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4" Dec 05 11:02:30.753160 master-0 kubenswrapper[24928]: I1205 11:02:30.752022 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdrqs\" (UniqueName: \"kubernetes.io/projected/be90920f-159a-4fac-9746-70846d306f22-kube-api-access-pdrqs\") pod \"horizon-operator-controller-manager-f6cc97788-5lr6c\" (UID: \"be90920f-159a-4fac-9746-70846d306f22\") " pod="openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c" Dec 05 11:02:30.784170 master-0 kubenswrapper[24928]: I1205 11:02:30.774385 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z"] Dec 05 11:02:30.805053 master-0 kubenswrapper[24928]: I1205 11:02:30.805003 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc" Dec 05 11:02:30.822913 master-0 kubenswrapper[24928]: I1205 11:02:30.817558 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4"] Dec 05 11:02:30.822913 master-0 kubenswrapper[24928]: I1205 11:02:30.817976 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:02:30.822913 master-0 kubenswrapper[24928]: I1205 11:02:30.818073 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brx8s\" (UniqueName: \"kubernetes.io/projected/da186c71-5e15-45a3-ad76-e51aad43439f-kube-api-access-brx8s\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:02:30.822913 master-0 kubenswrapper[24928]: I1205 11:02:30.818109 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjq2r\" (UniqueName: \"kubernetes.io/projected/3d85e846-d44c-49aa-8128-02566193d999-kube-api-access-jjq2r\") pod \"ironic-operator-controller-manager-7c9bfd6967-bhx8z\" (UID: \"3d85e846-d44c-49aa-8128-02566193d999\") " pod="openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z" Dec 05 11:02:30.822913 master-0 kubenswrapper[24928]: I1205 11:02:30.818142 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slkr9\" (UniqueName: \"kubernetes.io/projected/ada63c1e-313b-465a-a556-32b452de65ab-kube-api-access-slkr9\") pod \"keystone-operator-controller-manager-58b8dcc5fb-vv6s4\" (UID: \"ada63c1e-313b-465a-a556-32b452de65ab\") " pod="openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4" Dec 05 11:02:30.822913 master-0 kubenswrapper[24928]: E1205 11:02:30.818287 24928 secret.go:189] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:30.822913 master-0 kubenswrapper[24928]: E1205 11:02:30.818338 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert podName:da186c71-5e15-45a3-ad76-e51aad43439f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:31.318317312 +0000 UTC m=+911.321511163 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert") pod "infra-operator-controller-manager-7d9c9d7fd8-4ht2g" (UID: "da186c71-5e15-45a3-ad76-e51aad43439f") : secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:30.827191 master-0 kubenswrapper[24928]: I1205 11:02:30.827146 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc"] Dec 05 11:02:30.828571 master-0 kubenswrapper[24928]: I1205 11:02:30.828543 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc" Dec 05 11:02:30.845838 master-0 kubenswrapper[24928]: I1205 11:02:30.842318 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w"] Dec 05 11:02:30.845838 master-0 kubenswrapper[24928]: I1205 11:02:30.844052 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" Dec 05 11:02:30.848913 master-0 kubenswrapper[24928]: I1205 11:02:30.848680 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brx8s\" (UniqueName: \"kubernetes.io/projected/da186c71-5e15-45a3-ad76-e51aad43439f-kube-api-access-brx8s\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:02:30.852302 master-0 kubenswrapper[24928]: I1205 11:02:30.851913 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc"] Dec 05 11:02:30.885449 master-0 kubenswrapper[24928]: I1205 11:02:30.885062 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w"] Dec 05 11:02:30.895529 master-0 kubenswrapper[24928]: I1205 11:02:30.895441 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb"] Dec 05 11:02:30.911013 master-0 kubenswrapper[24928]: I1205 11:02:30.898968 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb" Dec 05 11:02:30.911405 master-0 kubenswrapper[24928]: I1205 11:02:30.911365 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6"] Dec 05 11:02:30.915398 master-0 kubenswrapper[24928]: I1205 11:02:30.912747 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" Dec 05 11:02:30.923788 master-0 kubenswrapper[24928]: I1205 11:02:30.922658 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vmh9\" (UniqueName: \"kubernetes.io/projected/442a2392-c758-4713-baa3-3e119ade664c-kube-api-access-4vmh9\") pod \"mariadb-operator-controller-manager-647d75769b-dft2w\" (UID: \"442a2392-c758-4713-baa3-3e119ade664c\") " pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" Dec 05 11:02:30.923788 master-0 kubenswrapper[24928]: I1205 11:02:30.922756 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjq2r\" (UniqueName: \"kubernetes.io/projected/3d85e846-d44c-49aa-8128-02566193d999-kube-api-access-jjq2r\") pod \"ironic-operator-controller-manager-7c9bfd6967-bhx8z\" (UID: \"3d85e846-d44c-49aa-8128-02566193d999\") " pod="openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z" Dec 05 11:02:30.923788 master-0 kubenswrapper[24928]: I1205 11:02:30.922813 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slkr9\" (UniqueName: \"kubernetes.io/projected/ada63c1e-313b-465a-a556-32b452de65ab-kube-api-access-slkr9\") pod \"keystone-operator-controller-manager-58b8dcc5fb-vv6s4\" (UID: \"ada63c1e-313b-465a-a556-32b452de65ab\") " pod="openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4" Dec 05 11:02:30.923788 master-0 kubenswrapper[24928]: I1205 11:02:30.923022 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjlhg\" (UniqueName: \"kubernetes.io/projected/61b1765d-2b89-4d57-85c0-a7f9afe632fe-kube-api-access-rjlhg\") pod \"manila-operator-controller-manager-56f9fbf74b-pwlgc\" (UID: \"61b1765d-2b89-4d57-85c0-a7f9afe632fe\") " pod="openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc" Dec 05 11:02:30.924113 master-0 kubenswrapper[24928]: I1205 11:02:30.922638 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb"] Dec 05 11:02:30.933939 master-0 kubenswrapper[24928]: I1205 11:02:30.933872 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6"] Dec 05 11:02:30.939095 master-0 kubenswrapper[24928]: I1205 11:02:30.936010 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc" Dec 05 11:02:30.960631 master-0 kubenswrapper[24928]: I1205 11:02:30.953463 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slkr9\" (UniqueName: \"kubernetes.io/projected/ada63c1e-313b-465a-a556-32b452de65ab-kube-api-access-slkr9\") pod \"keystone-operator-controller-manager-58b8dcc5fb-vv6s4\" (UID: \"ada63c1e-313b-465a-a556-32b452de65ab\") " pod="openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4" Dec 05 11:02:30.960631 master-0 kubenswrapper[24928]: I1205 11:02:30.956112 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjq2r\" (UniqueName: \"kubernetes.io/projected/3d85e846-d44c-49aa-8128-02566193d999-kube-api-access-jjq2r\") pod \"ironic-operator-controller-manager-7c9bfd6967-bhx8z\" (UID: \"3d85e846-d44c-49aa-8128-02566193d999\") " pod="openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z" Dec 05 11:02:30.993080 master-0 kubenswrapper[24928]: I1205 11:02:30.992202 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c" Dec 05 11:02:30.998788 master-0 kubenswrapper[24928]: I1205 11:02:30.997926 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls"] Dec 05 11:02:31.000261 master-0 kubenswrapper[24928]: I1205 11:02:31.000207 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls" Dec 05 11:02:31.017852 master-0 kubenswrapper[24928]: I1205 11:02:31.017800 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls"] Dec 05 11:02:31.025754 master-0 kubenswrapper[24928]: I1205 11:02:31.025477 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjlhg\" (UniqueName: \"kubernetes.io/projected/61b1765d-2b89-4d57-85c0-a7f9afe632fe-kube-api-access-rjlhg\") pod \"manila-operator-controller-manager-56f9fbf74b-pwlgc\" (UID: \"61b1765d-2b89-4d57-85c0-a7f9afe632fe\") " pod="openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc" Dec 05 11:02:31.025754 master-0 kubenswrapper[24928]: I1205 11:02:31.025605 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn6c6\" (UniqueName: \"kubernetes.io/projected/5af60df4-8ca0-4bfb-8302-d3f9dcd4ef7a-kube-api-access-tn6c6\") pod \"neutron-operator-controller-manager-7cdd6b54fb-9wfjb\" (UID: \"5af60df4-8ca0-4bfb-8302-d3f9dcd4ef7a\") " pod="openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb" Dec 05 11:02:31.025754 master-0 kubenswrapper[24928]: I1205 11:02:31.025651 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vmh9\" (UniqueName: \"kubernetes.io/projected/442a2392-c758-4713-baa3-3e119ade664c-kube-api-access-4vmh9\") pod \"mariadb-operator-controller-manager-647d75769b-dft2w\" (UID: \"442a2392-c758-4713-baa3-3e119ade664c\") " pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" Dec 05 11:02:31.025754 master-0 kubenswrapper[24928]: I1205 11:02:31.025710 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srtkc\" (UniqueName: \"kubernetes.io/projected/fa9681c6-d40e-4075-9230-1ed7431a06fb-kube-api-access-srtkc\") pod \"nova-operator-controller-manager-865fc86d5b-z8jv6\" (UID: \"fa9681c6-d40e-4075-9230-1ed7431a06fb\") " pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" Dec 05 11:02:31.050836 master-0 kubenswrapper[24928]: I1205 11:02:31.050781 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjlhg\" (UniqueName: \"kubernetes.io/projected/61b1765d-2b89-4d57-85c0-a7f9afe632fe-kube-api-access-rjlhg\") pod \"manila-operator-controller-manager-56f9fbf74b-pwlgc\" (UID: \"61b1765d-2b89-4d57-85c0-a7f9afe632fe\") " pod="openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc" Dec 05 11:02:31.052467 master-0 kubenswrapper[24928]: I1205 11:02:31.052391 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w"] Dec 05 11:02:31.055124 master-0 kubenswrapper[24928]: I1205 11:02:31.054278 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" Dec 05 11:02:31.059384 master-0 kubenswrapper[24928]: I1205 11:02:31.059336 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vmh9\" (UniqueName: \"kubernetes.io/projected/442a2392-c758-4713-baa3-3e119ade664c-kube-api-access-4vmh9\") pod \"mariadb-operator-controller-manager-647d75769b-dft2w\" (UID: \"442a2392-c758-4713-baa3-3e119ade664c\") " pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" Dec 05 11:02:31.086687 master-0 kubenswrapper[24928]: I1205 11:02:31.086599 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w"] Dec 05 11:02:31.098486 master-0 kubenswrapper[24928]: I1205 11:02:31.093445 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:02:31.098486 master-0 kubenswrapper[24928]: I1205 11:02:31.095295 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 05 11:02:31.121530 master-0 kubenswrapper[24928]: I1205 11:02:31.116579 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w"] Dec 05 11:02:31.124766 master-0 kubenswrapper[24928]: I1205 11:02:31.123028 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z" Dec 05 11:02:31.126821 master-0 kubenswrapper[24928]: I1205 11:02:31.126754 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srtkc\" (UniqueName: \"kubernetes.io/projected/fa9681c6-d40e-4075-9230-1ed7431a06fb-kube-api-access-srtkc\") pod \"nova-operator-controller-manager-865fc86d5b-z8jv6\" (UID: \"fa9681c6-d40e-4075-9230-1ed7431a06fb\") " pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" Dec 05 11:02:31.126908 master-0 kubenswrapper[24928]: I1205 11:02:31.126887 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jktl9\" (UniqueName: \"kubernetes.io/projected/da92e42f-68d5-4f34-bcf5-c5a8e765e383-kube-api-access-jktl9\") pod \"octavia-operator-controller-manager-845b79dc4f-dc9ls\" (UID: \"da92e42f-68d5-4f34-bcf5-c5a8e765e383\") " pod="openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls" Dec 05 11:02:31.126976 master-0 kubenswrapper[24928]: I1205 11:02:31.126924 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn6c6\" (UniqueName: \"kubernetes.io/projected/5af60df4-8ca0-4bfb-8302-d3f9dcd4ef7a-kube-api-access-tn6c6\") pod \"neutron-operator-controller-manager-7cdd6b54fb-9wfjb\" (UID: \"5af60df4-8ca0-4bfb-8302-d3f9dcd4ef7a\") " pod="openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb" Dec 05 11:02:31.126976 master-0 kubenswrapper[24928]: I1205 11:02:31.126948 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgs2h\" (UniqueName: \"kubernetes.io/projected/68a107e3-b0eb-43fd-844f-817cf01ef767-kube-api-access-tgs2h\") pod \"ovn-operator-controller-manager-647f96877-gcg9w\" (UID: \"68a107e3-b0eb-43fd-844f-817cf01ef767\") " pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" Dec 05 11:02:31.139325 master-0 kubenswrapper[24928]: I1205 11:02:31.139267 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4" Dec 05 11:02:31.143269 master-0 kubenswrapper[24928]: I1205 11:02:31.143217 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w"] Dec 05 11:02:31.154217 master-0 kubenswrapper[24928]: I1205 11:02:31.153248 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs"] Dec 05 11:02:31.156132 master-0 kubenswrapper[24928]: I1205 11:02:31.155100 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" Dec 05 11:02:31.162768 master-0 kubenswrapper[24928]: I1205 11:02:31.162718 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn6c6\" (UniqueName: \"kubernetes.io/projected/5af60df4-8ca0-4bfb-8302-d3f9dcd4ef7a-kube-api-access-tn6c6\") pod \"neutron-operator-controller-manager-7cdd6b54fb-9wfjb\" (UID: \"5af60df4-8ca0-4bfb-8302-d3f9dcd4ef7a\") " pod="openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb" Dec 05 11:02:31.165547 master-0 kubenswrapper[24928]: I1205 11:02:31.162905 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-696b999796-bwcl8"] Dec 05 11:02:31.169159 master-0 kubenswrapper[24928]: I1205 11:02:31.169101 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-696b999796-bwcl8" Dec 05 11:02:31.171232 master-0 kubenswrapper[24928]: I1205 11:02:31.169642 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc" Dec 05 11:02:31.171232 master-0 kubenswrapper[24928]: I1205 11:02:31.170149 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs"] Dec 05 11:02:31.206278 master-0 kubenswrapper[24928]: I1205 11:02:31.197046 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srtkc\" (UniqueName: \"kubernetes.io/projected/fa9681c6-d40e-4075-9230-1ed7431a06fb-kube-api-access-srtkc\") pod \"nova-operator-controller-manager-865fc86d5b-z8jv6\" (UID: \"fa9681c6-d40e-4075-9230-1ed7431a06fb\") " pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" Dec 05 11:02:31.206278 master-0 kubenswrapper[24928]: I1205 11:02:31.199634 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" Dec 05 11:02:31.219535 master-0 kubenswrapper[24928]: I1205 11:02:31.217509 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-696b999796-bwcl8"] Dec 05 11:02:31.224740 master-0 kubenswrapper[24928]: I1205 11:02:31.224672 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb" Dec 05 11:02:31.251578 master-0 kubenswrapper[24928]: I1205 11:02:31.245388 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" Dec 05 11:02:31.251578 master-0 kubenswrapper[24928]: I1205 11:02:31.247522 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgs2h\" (UniqueName: \"kubernetes.io/projected/68a107e3-b0eb-43fd-844f-817cf01ef767-kube-api-access-tgs2h\") pod \"ovn-operator-controller-manager-647f96877-gcg9w\" (UID: \"68a107e3-b0eb-43fd-844f-817cf01ef767\") " pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" Dec 05 11:02:31.251578 master-0 kubenswrapper[24928]: I1205 11:02:31.247581 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl74s\" (UniqueName: \"kubernetes.io/projected/cd637a6d-86b1-4854-84d5-a86c73fda74a-kube-api-access-bl74s\") pod \"swift-operator-controller-manager-696b999796-bwcl8\" (UID: \"cd637a6d-86b1-4854-84d5-a86c73fda74a\") " pod="openstack-operators/swift-operator-controller-manager-696b999796-bwcl8" Dec 05 11:02:31.251578 master-0 kubenswrapper[24928]: I1205 11:02:31.247674 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtnmg\" (UniqueName: \"kubernetes.io/projected/5b821d96-dc12-4794-a863-0cc47d4d8d65-kube-api-access-gtnmg\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:02:31.251578 master-0 kubenswrapper[24928]: I1205 11:02:31.247715 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:02:31.251578 master-0 kubenswrapper[24928]: I1205 11:02:31.247810 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jktl9\" (UniqueName: \"kubernetes.io/projected/da92e42f-68d5-4f34-bcf5-c5a8e765e383-kube-api-access-jktl9\") pod \"octavia-operator-controller-manager-845b79dc4f-dc9ls\" (UID: \"da92e42f-68d5-4f34-bcf5-c5a8e765e383\") " pod="openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls" Dec 05 11:02:31.251578 master-0 kubenswrapper[24928]: I1205 11:02:31.247835 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v26t2\" (UniqueName: \"kubernetes.io/projected/a4e20703-5c0d-44a1-94c6-f5b97782ff5c-kube-api-access-v26t2\") pod \"placement-operator-controller-manager-6b64f6f645-xf7hs\" (UID: \"a4e20703-5c0d-44a1-94c6-f5b97782ff5c\") " pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" Dec 05 11:02:31.260475 master-0 kubenswrapper[24928]: I1205 11:02:31.260281 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4"] Dec 05 11:02:31.263840 master-0 kubenswrapper[24928]: I1205 11:02:31.263797 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4" Dec 05 11:02:31.266553 master-0 kubenswrapper[24928]: I1205 11:02:31.265698 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4"] Dec 05 11:02:31.267332 master-0 kubenswrapper[24928]: I1205 11:02:31.267284 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jktl9\" (UniqueName: \"kubernetes.io/projected/da92e42f-68d5-4f34-bcf5-c5a8e765e383-kube-api-access-jktl9\") pod \"octavia-operator-controller-manager-845b79dc4f-dc9ls\" (UID: \"da92e42f-68d5-4f34-bcf5-c5a8e765e383\") " pod="openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls" Dec 05 11:02:31.274019 master-0 kubenswrapper[24928]: I1205 11:02:31.273961 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgs2h\" (UniqueName: \"kubernetes.io/projected/68a107e3-b0eb-43fd-844f-817cf01ef767-kube-api-access-tgs2h\") pod \"ovn-operator-controller-manager-647f96877-gcg9w\" (UID: \"68a107e3-b0eb-43fd-844f-817cf01ef767\") " pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" Dec 05 11:02:31.287132 master-0 kubenswrapper[24928]: I1205 11:02:31.287076 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m"] Dec 05 11:02:31.288932 master-0 kubenswrapper[24928]: I1205 11:02:31.288839 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" Dec 05 11:02:31.317712 master-0 kubenswrapper[24928]: I1205 11:02:31.314107 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m"] Dec 05 11:02:31.318240 master-0 kubenswrapper[24928]: I1205 11:02:31.318051 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls" Dec 05 11:02:31.345515 master-0 kubenswrapper[24928]: I1205 11:02:31.345476 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b"] Dec 05 11:02:31.347281 master-0 kubenswrapper[24928]: I1205 11:02:31.347235 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" Dec 05 11:02:31.351221 master-0 kubenswrapper[24928]: I1205 11:02:31.349115 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtnmg\" (UniqueName: \"kubernetes.io/projected/5b821d96-dc12-4794-a863-0cc47d4d8d65-kube-api-access-gtnmg\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:02:31.351221 master-0 kubenswrapper[24928]: I1205 11:02:31.349161 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:02:31.351221 master-0 kubenswrapper[24928]: I1205 11:02:31.349194 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jp5k9\" (UniqueName: \"kubernetes.io/projected/7cc35a07-e403-442b-bcdf-836525eff1f7-kube-api-access-jp5k9\") pod \"test-operator-controller-manager-57dfcdd5b8-rth9m\" (UID: \"7cc35a07-e403-442b-bcdf-836525eff1f7\") " pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" Dec 05 11:02:31.351221 master-0 kubenswrapper[24928]: I1205 11:02:31.349249 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:02:31.351221 master-0 kubenswrapper[24928]: I1205 11:02:31.349273 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v26t2\" (UniqueName: \"kubernetes.io/projected/a4e20703-5c0d-44a1-94c6-f5b97782ff5c-kube-api-access-v26t2\") pod \"placement-operator-controller-manager-6b64f6f645-xf7hs\" (UID: \"a4e20703-5c0d-44a1-94c6-f5b97782ff5c\") " pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" Dec 05 11:02:31.351221 master-0 kubenswrapper[24928]: I1205 11:02:31.349302 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl74s\" (UniqueName: \"kubernetes.io/projected/cd637a6d-86b1-4854-84d5-a86c73fda74a-kube-api-access-bl74s\") pod \"swift-operator-controller-manager-696b999796-bwcl8\" (UID: \"cd637a6d-86b1-4854-84d5-a86c73fda74a\") " pod="openstack-operators/swift-operator-controller-manager-696b999796-bwcl8" Dec 05 11:02:31.351221 master-0 kubenswrapper[24928]: I1205 11:02:31.349325 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxkjc\" (UniqueName: \"kubernetes.io/projected/16b00867-62d3-469b-b9c1-98c2e8a6bf47-kube-api-access-jxkjc\") pod \"telemetry-operator-controller-manager-7b5867bfc7-7gjc4\" (UID: \"16b00867-62d3-469b-b9c1-98c2e8a6bf47\") " pod="openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4" Dec 05 11:02:31.359032 master-0 kubenswrapper[24928]: E1205 11:02:31.358049 24928 secret.go:189] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:31.359032 master-0 kubenswrapper[24928]: E1205 11:02:31.358113 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert podName:da186c71-5e15-45a3-ad76-e51aad43439f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:32.358097854 +0000 UTC m=+912.361291705 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert") pod "infra-operator-controller-manager-7d9c9d7fd8-4ht2g" (UID: "da186c71-5e15-45a3-ad76-e51aad43439f") : secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:31.359032 master-0 kubenswrapper[24928]: E1205 11:02:31.358474 24928 secret.go:189] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:31.359032 master-0 kubenswrapper[24928]: E1205 11:02:31.358501 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert podName:5b821d96-dc12-4794-a863-0cc47d4d8d65 nodeName:}" failed. No retries permitted until 2025-12-05 11:02:31.858493715 +0000 UTC m=+911.861687566 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert") pod "openstack-baremetal-operator-controller-manager-6f998f574688x6w" (UID: "5b821d96-dc12-4794-a863-0cc47d4d8d65") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:31.388494 master-0 kubenswrapper[24928]: I1205 11:02:31.377029 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b"] Dec 05 11:02:31.388772 master-0 kubenswrapper[24928]: I1205 11:02:31.388733 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" Dec 05 11:02:31.390563 master-0 kubenswrapper[24928]: I1205 11:02:31.390522 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl74s\" (UniqueName: \"kubernetes.io/projected/cd637a6d-86b1-4854-84d5-a86c73fda74a-kube-api-access-bl74s\") pod \"swift-operator-controller-manager-696b999796-bwcl8\" (UID: \"cd637a6d-86b1-4854-84d5-a86c73fda74a\") " pod="openstack-operators/swift-operator-controller-manager-696b999796-bwcl8" Dec 05 11:02:31.398795 master-0 kubenswrapper[24928]: I1205 11:02:31.398756 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v26t2\" (UniqueName: \"kubernetes.io/projected/a4e20703-5c0d-44a1-94c6-f5b97782ff5c-kube-api-access-v26t2\") pod \"placement-operator-controller-manager-6b64f6f645-xf7hs\" (UID: \"a4e20703-5c0d-44a1-94c6-f5b97782ff5c\") " pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" Dec 05 11:02:31.405518 master-0 kubenswrapper[24928]: I1205 11:02:31.405437 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtnmg\" (UniqueName: \"kubernetes.io/projected/5b821d96-dc12-4794-a863-0cc47d4d8d65-kube-api-access-gtnmg\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:02:31.444930 master-0 kubenswrapper[24928]: I1205 11:02:31.444866 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692"] Dec 05 11:02:31.446534 master-0 kubenswrapper[24928]: I1205 11:02:31.446495 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:31.450328 master-0 kubenswrapper[24928]: I1205 11:02:31.450278 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 05 11:02:31.450723 master-0 kubenswrapper[24928]: I1205 11:02:31.450694 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 05 11:02:31.451610 master-0 kubenswrapper[24928]: I1205 11:02:31.450905 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxkjc\" (UniqueName: \"kubernetes.io/projected/16b00867-62d3-469b-b9c1-98c2e8a6bf47-kube-api-access-jxkjc\") pod \"telemetry-operator-controller-manager-7b5867bfc7-7gjc4\" (UID: \"16b00867-62d3-469b-b9c1-98c2e8a6bf47\") " pod="openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4" Dec 05 11:02:31.451610 master-0 kubenswrapper[24928]: I1205 11:02:31.451016 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pdcx\" (UniqueName: \"kubernetes.io/projected/acbb966c-9984-42a4-9739-1cf87d7db982-kube-api-access-4pdcx\") pod \"watcher-operator-controller-manager-6b9b669fdb-tsk7b\" (UID: \"acbb966c-9984-42a4-9739-1cf87d7db982\") " pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" Dec 05 11:02:31.451610 master-0 kubenswrapper[24928]: I1205 11:02:31.451056 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jp5k9\" (UniqueName: \"kubernetes.io/projected/7cc35a07-e403-442b-bcdf-836525eff1f7-kube-api-access-jp5k9\") pod \"test-operator-controller-manager-57dfcdd5b8-rth9m\" (UID: \"7cc35a07-e403-442b-bcdf-836525eff1f7\") " pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" Dec 05 11:02:31.458956 master-0 kubenswrapper[24928]: I1205 11:02:31.456755 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692"] Dec 05 11:02:31.479325 master-0 kubenswrapper[24928]: I1205 11:02:31.479277 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxkjc\" (UniqueName: \"kubernetes.io/projected/16b00867-62d3-469b-b9c1-98c2e8a6bf47-kube-api-access-jxkjc\") pod \"telemetry-operator-controller-manager-7b5867bfc7-7gjc4\" (UID: \"16b00867-62d3-469b-b9c1-98c2e8a6bf47\") " pod="openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4" Dec 05 11:02:31.483502 master-0 kubenswrapper[24928]: I1205 11:02:31.481249 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jp5k9\" (UniqueName: \"kubernetes.io/projected/7cc35a07-e403-442b-bcdf-836525eff1f7-kube-api-access-jp5k9\") pod \"test-operator-controller-manager-57dfcdd5b8-rth9m\" (UID: \"7cc35a07-e403-442b-bcdf-836525eff1f7\") " pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" Dec 05 11:02:31.498703 master-0 kubenswrapper[24928]: I1205 11:02:31.496892 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-78955d896f-8fcxk"] Dec 05 11:02:31.505302 master-0 kubenswrapper[24928]: I1205 11:02:31.503991 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-78955d896f-8fcxk" Dec 05 11:02:31.509531 master-0 kubenswrapper[24928]: I1205 11:02:31.508989 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-78955d896f-8fcxk"] Dec 05 11:02:31.518773 master-0 kubenswrapper[24928]: I1205 11:02:31.518469 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" Dec 05 11:02:31.531206 master-0 kubenswrapper[24928]: I1205 11:02:31.531145 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v"] Dec 05 11:02:31.554577 master-0 kubenswrapper[24928]: I1205 11:02:31.553348 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:31.554577 master-0 kubenswrapper[24928]: I1205 11:02:31.553487 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thq9j\" (UniqueName: \"kubernetes.io/projected/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-kube-api-access-thq9j\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:31.554577 master-0 kubenswrapper[24928]: I1205 11:02:31.553546 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:31.554577 master-0 kubenswrapper[24928]: I1205 11:02:31.553634 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pdcx\" (UniqueName: \"kubernetes.io/projected/acbb966c-9984-42a4-9739-1cf87d7db982-kube-api-access-4pdcx\") pod \"watcher-operator-controller-manager-6b9b669fdb-tsk7b\" (UID: \"acbb966c-9984-42a4-9739-1cf87d7db982\") " pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" Dec 05 11:02:31.554577 master-0 kubenswrapper[24928]: I1205 11:02:31.553666 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lhvc\" (UniqueName: \"kubernetes.io/projected/5db6830c-28ff-4fe3-bf7e-1f417df8bfeb-kube-api-access-4lhvc\") pod \"rabbitmq-cluster-operator-manager-78955d896f-8fcxk\" (UID: \"5db6830c-28ff-4fe3-bf7e-1f417df8bfeb\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-78955d896f-8fcxk" Dec 05 11:02:31.568692 master-0 kubenswrapper[24928]: I1205 11:02:31.568468 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v" event={"ID":"4f712f48-af26-4222-a231-ac989f78d18f","Type":"ContainerStarted","Data":"5bbdb15ea87075a1f420ab1a6024c0297aa983718418271b7e907bf7b8052480"} Dec 05 11:02:31.586229 master-0 kubenswrapper[24928]: I1205 11:02:31.585106 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pdcx\" (UniqueName: \"kubernetes.io/projected/acbb966c-9984-42a4-9739-1cf87d7db982-kube-api-access-4pdcx\") pod \"watcher-operator-controller-manager-6b9b669fdb-tsk7b\" (UID: \"acbb966c-9984-42a4-9739-1cf87d7db982\") " pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" Dec 05 11:02:31.595062 master-0 kubenswrapper[24928]: I1205 11:02:31.594109 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-696b999796-bwcl8" Dec 05 11:02:31.615066 master-0 kubenswrapper[24928]: I1205 11:02:31.615007 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4" Dec 05 11:02:31.641723 master-0 kubenswrapper[24928]: I1205 11:02:31.641197 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" Dec 05 11:02:31.655778 master-0 kubenswrapper[24928]: I1205 11:02:31.655705 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:31.655952 master-0 kubenswrapper[24928]: I1205 11:02:31.655827 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lhvc\" (UniqueName: \"kubernetes.io/projected/5db6830c-28ff-4fe3-bf7e-1f417df8bfeb-kube-api-access-4lhvc\") pod \"rabbitmq-cluster-operator-manager-78955d896f-8fcxk\" (UID: \"5db6830c-28ff-4fe3-bf7e-1f417df8bfeb\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-78955d896f-8fcxk" Dec 05 11:02:31.655991 master-0 kubenswrapper[24928]: I1205 11:02:31.655965 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:31.656538 master-0 kubenswrapper[24928]: I1205 11:02:31.656018 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thq9j\" (UniqueName: \"kubernetes.io/projected/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-kube-api-access-thq9j\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:31.656538 master-0 kubenswrapper[24928]: E1205 11:02:31.656510 24928 secret.go:189] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 11:02:31.656630 master-0 kubenswrapper[24928]: E1205 11:02:31.656568 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:32.156546659 +0000 UTC m=+912.159740520 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "metrics-server-cert" not found Dec 05 11:02:31.656998 master-0 kubenswrapper[24928]: E1205 11:02:31.656799 24928 secret.go:189] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 11:02:31.656998 master-0 kubenswrapper[24928]: E1205 11:02:31.656839 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:32.156828186 +0000 UTC m=+912.160022037 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "webhook-server-cert" not found Dec 05 11:02:31.675723 master-0 kubenswrapper[24928]: I1205 11:02:31.673929 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lhvc\" (UniqueName: \"kubernetes.io/projected/5db6830c-28ff-4fe3-bf7e-1f417df8bfeb-kube-api-access-4lhvc\") pod \"rabbitmq-cluster-operator-manager-78955d896f-8fcxk\" (UID: \"5db6830c-28ff-4fe3-bf7e-1f417df8bfeb\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-78955d896f-8fcxk" Dec 05 11:02:31.678286 master-0 kubenswrapper[24928]: I1205 11:02:31.678234 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thq9j\" (UniqueName: \"kubernetes.io/projected/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-kube-api-access-thq9j\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:31.732064 master-0 kubenswrapper[24928]: I1205 11:02:31.731952 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" Dec 05 11:02:31.781290 master-0 kubenswrapper[24928]: I1205 11:02:31.781227 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc"] Dec 05 11:02:31.801422 master-0 kubenswrapper[24928]: I1205 11:02:31.801322 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2"] Dec 05 11:02:31.804435 master-0 kubenswrapper[24928]: W1205 11:02:31.804292 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5017646a_fb39_440c_afe5_a4d6cee68ed4.slice/crio-16448c977afb1702284be4832b4c8a832bd6da09c9bdbc579dc657ff012287af WatchSource:0}: Error finding container 16448c977afb1702284be4832b4c8a832bd6da09c9bdbc579dc657ff012287af: Status 404 returned error can't find the container with id 16448c977afb1702284be4832b4c8a832bd6da09c9bdbc579dc657ff012287af Dec 05 11:02:31.828558 master-0 kubenswrapper[24928]: I1205 11:02:31.828386 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w"] Dec 05 11:02:31.840950 master-0 kubenswrapper[24928]: I1205 11:02:31.840372 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-78955d896f-8fcxk" Dec 05 11:02:31.859693 master-0 kubenswrapper[24928]: I1205 11:02:31.859619 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:02:31.860142 master-0 kubenswrapper[24928]: E1205 11:02:31.859990 24928 secret.go:189] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:31.860142 master-0 kubenswrapper[24928]: E1205 11:02:31.860059 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert podName:5b821d96-dc12-4794-a863-0cc47d4d8d65 nodeName:}" failed. No retries permitted until 2025-12-05 11:02:32.860040044 +0000 UTC m=+912.863233895 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert") pod "openstack-baremetal-operator-controller-manager-6f998f574688x6w" (UID: "5b821d96-dc12-4794-a863-0cc47d4d8d65") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:31.862248 master-0 kubenswrapper[24928]: I1205 11:02:31.862214 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c"] Dec 05 11:02:31.874060 master-0 kubenswrapper[24928]: I1205 11:02:31.873964 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc"] Dec 05 11:02:32.193668 master-0 kubenswrapper[24928]: I1205 11:02:32.193588 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:32.193877 master-0 kubenswrapper[24928]: E1205 11:02:32.193811 24928 secret.go:189] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 11:02:32.193951 master-0 kubenswrapper[24928]: E1205 11:02:32.193914 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:33.19389108 +0000 UTC m=+913.197084931 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "metrics-server-cert" not found Dec 05 11:02:32.194109 master-0 kubenswrapper[24928]: I1205 11:02:32.194073 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:32.194347 master-0 kubenswrapper[24928]: E1205 11:02:32.194301 24928 secret.go:189] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 11:02:32.194440 master-0 kubenswrapper[24928]: E1205 11:02:32.194407 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:33.194382983 +0000 UTC m=+913.197576834 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "webhook-server-cert" not found Dec 05 11:02:32.283389 master-0 kubenswrapper[24928]: I1205 11:02:32.283319 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4"] Dec 05 11:02:32.292741 master-0 kubenswrapper[24928]: I1205 11:02:32.292680 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc"] Dec 05 11:02:32.308816 master-0 kubenswrapper[24928]: W1205 11:02:32.308327 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod442a2392_c758_4713_baa3_3e119ade664c.slice/crio-f4d29a6333f54d3ddc952106705183048dfac2977903940099a871d20d27b040 WatchSource:0}: Error finding container f4d29a6333f54d3ddc952106705183048dfac2977903940099a871d20d27b040: Status 404 returned error can't find the container with id f4d29a6333f54d3ddc952106705183048dfac2977903940099a871d20d27b040 Dec 05 11:02:32.316267 master-0 kubenswrapper[24928]: I1205 11:02:32.316205 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w"] Dec 05 11:02:32.318596 master-0 kubenswrapper[24928]: W1205 11:02:32.318551 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d85e846_d44c_49aa_8128_02566193d999.slice/crio-90c4200631254f8d3ea04b8be3cc4529b2253d2c98e2659e0e23e82436ef6609 WatchSource:0}: Error finding container 90c4200631254f8d3ea04b8be3cc4529b2253d2c98e2659e0e23e82436ef6609: Status 404 returned error can't find the container with id 90c4200631254f8d3ea04b8be3cc4529b2253d2c98e2659e0e23e82436ef6609 Dec 05 11:02:32.327946 master-0 kubenswrapper[24928]: I1205 11:02:32.327879 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z"] Dec 05 11:02:32.398222 master-0 kubenswrapper[24928]: I1205 11:02:32.398142 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:02:32.398455 master-0 kubenswrapper[24928]: E1205 11:02:32.398331 24928 secret.go:189] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:32.398455 master-0 kubenswrapper[24928]: E1205 11:02:32.398413 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert podName:da186c71-5e15-45a3-ad76-e51aad43439f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:34.39839055 +0000 UTC m=+914.401584401 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert") pod "infra-operator-controller-manager-7d9c9d7fd8-4ht2g" (UID: "da186c71-5e15-45a3-ad76-e51aad43439f") : secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:32.581479 master-0 kubenswrapper[24928]: I1205 11:02:32.580839 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z" event={"ID":"3d85e846-d44c-49aa-8128-02566193d999","Type":"ContainerStarted","Data":"90c4200631254f8d3ea04b8be3cc4529b2253d2c98e2659e0e23e82436ef6609"} Dec 05 11:02:32.582213 master-0 kubenswrapper[24928]: I1205 11:02:32.582150 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc" event={"ID":"61b1765d-2b89-4d57-85c0-a7f9afe632fe","Type":"ContainerStarted","Data":"0487a50e4b1b1af0ad34bf1608a01e124a22f244789933b48181f022f317e2a9"} Dec 05 11:02:32.583100 master-0 kubenswrapper[24928]: I1205 11:02:32.583061 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc" event={"ID":"216bdcf3-b763-4293-8ac4-3e7eeeae0b5d","Type":"ContainerStarted","Data":"7af51dad2966ce302aa9315658793548e1c8793e76ef38833e79d830ba812d9e"} Dec 05 11:02:32.583969 master-0 kubenswrapper[24928]: I1205 11:02:32.583932 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4" event={"ID":"ada63c1e-313b-465a-a556-32b452de65ab","Type":"ContainerStarted","Data":"51f0fcc561b43234b3974b18d74a7d5e601de29998717900d5648eaab1b43fe3"} Dec 05 11:02:32.585481 master-0 kubenswrapper[24928]: I1205 11:02:32.585443 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" event={"ID":"442a2392-c758-4713-baa3-3e119ade664c","Type":"ContainerStarted","Data":"f4d29a6333f54d3ddc952106705183048dfac2977903940099a871d20d27b040"} Dec 05 11:02:32.587683 master-0 kubenswrapper[24928]: I1205 11:02:32.587617 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2" event={"ID":"5017646a-fb39-440c-afe5-a4d6cee68ed4","Type":"ContainerStarted","Data":"16448c977afb1702284be4832b4c8a832bd6da09c9bdbc579dc657ff012287af"} Dec 05 11:02:32.588778 master-0 kubenswrapper[24928]: I1205 11:02:32.588737 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w" event={"ID":"44343d25-46da-40a9-b0be-94861421d403","Type":"ContainerStarted","Data":"dadf580c831379bfabcdb119f902d03d929e21bf9c72cd07eab665ff61438681"} Dec 05 11:02:32.590182 master-0 kubenswrapper[24928]: I1205 11:02:32.590146 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c" event={"ID":"be90920f-159a-4fac-9746-70846d306f22","Type":"ContainerStarted","Data":"670a356fd577605f99a9aa00968a160c4283daf8875d85da2f6f80b06b0f4c47"} Dec 05 11:02:32.592265 master-0 kubenswrapper[24928]: I1205 11:02:32.592221 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc" event={"ID":"1d175522-0c08-40c5-a309-f1820406ea2f","Type":"ContainerStarted","Data":"733fda8c5b29b5f10901f8362594220ef25cf045fd0d92ebaa8097e44a37672a"} Dec 05 11:02:32.718775 master-0 kubenswrapper[24928]: I1205 11:02:32.718719 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs"] Dec 05 11:02:32.724188 master-0 kubenswrapper[24928]: W1205 11:02:32.724112 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4e20703_5c0d_44a1_94c6_f5b97782ff5c.slice/crio-ac9afeb11481e27bc7b7b6e4e84e5c947402e5c5f1ce16803078c606507ecf51 WatchSource:0}: Error finding container ac9afeb11481e27bc7b7b6e4e84e5c947402e5c5f1ce16803078c606507ecf51: Status 404 returned error can't find the container with id ac9afeb11481e27bc7b7b6e4e84e5c947402e5c5f1ce16803078c606507ecf51 Dec 05 11:02:32.744662 master-0 kubenswrapper[24928]: W1205 11:02:32.744618 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda92e42f_68d5_4f34_bcf5_c5a8e765e383.slice/crio-55d6b34ade76b9e8685e052d887d3efd3ffcd54d5c1e658fbf13cd48af4e6f92 WatchSource:0}: Error finding container 55d6b34ade76b9e8685e052d887d3efd3ffcd54d5c1e658fbf13cd48af4e6f92: Status 404 returned error can't find the container with id 55d6b34ade76b9e8685e052d887d3efd3ffcd54d5c1e658fbf13cd48af4e6f92 Dec 05 11:02:32.752998 master-0 kubenswrapper[24928]: I1205 11:02:32.752950 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb"] Dec 05 11:02:32.768670 master-0 kubenswrapper[24928]: I1205 11:02:32.768599 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls"] Dec 05 11:02:32.777493 master-0 kubenswrapper[24928]: I1205 11:02:32.777386 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w"] Dec 05 11:02:32.797971 master-0 kubenswrapper[24928]: I1205 11:02:32.797913 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6"] Dec 05 11:02:32.863331 master-0 kubenswrapper[24928]: I1205 11:02:32.863193 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:02:32.863578 master-0 kubenswrapper[24928]: E1205 11:02:32.863554 24928 secret.go:189] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:32.863641 master-0 kubenswrapper[24928]: E1205 11:02:32.863619 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert podName:5b821d96-dc12-4794-a863-0cc47d4d8d65 nodeName:}" failed. No retries permitted until 2025-12-05 11:02:34.863600514 +0000 UTC m=+914.866794365 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert") pod "openstack-baremetal-operator-controller-manager-6f998f574688x6w" (UID: "5b821d96-dc12-4794-a863-0cc47d4d8d65") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:33.184114 master-0 kubenswrapper[24928]: I1205 11:02:33.184042 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-696b999796-bwcl8"] Dec 05 11:02:33.200345 master-0 kubenswrapper[24928]: I1205 11:02:33.200261 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-78955d896f-8fcxk"] Dec 05 11:02:33.220151 master-0 kubenswrapper[24928]: I1205 11:02:33.219548 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m"] Dec 05 11:02:33.238574 master-0 kubenswrapper[24928]: I1205 11:02:33.238500 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b"] Dec 05 11:02:33.250492 master-0 kubenswrapper[24928]: I1205 11:02:33.250414 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4"] Dec 05 11:02:33.277010 master-0 kubenswrapper[24928]: I1205 11:02:33.276954 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:33.277193 master-0 kubenswrapper[24928]: I1205 11:02:33.277040 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:33.277193 master-0 kubenswrapper[24928]: E1205 11:02:33.277182 24928 secret.go:189] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 11:02:33.277272 master-0 kubenswrapper[24928]: E1205 11:02:33.277230 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:35.277215109 +0000 UTC m=+915.280408960 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "metrics-server-cert" not found Dec 05 11:02:33.277645 master-0 kubenswrapper[24928]: E1205 11:02:33.277602 24928 secret.go:189] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 11:02:33.277758 master-0 kubenswrapper[24928]: E1205 11:02:33.277732 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:35.277709021 +0000 UTC m=+915.280902932 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "webhook-server-cert" not found Dec 05 11:02:33.609913 master-0 kubenswrapper[24928]: I1205 11:02:33.609819 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" event={"ID":"a4e20703-5c0d-44a1-94c6-f5b97782ff5c","Type":"ContainerStarted","Data":"ac9afeb11481e27bc7b7b6e4e84e5c947402e5c5f1ce16803078c606507ecf51"} Dec 05 11:02:33.613018 master-0 kubenswrapper[24928]: I1205 11:02:33.612916 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls" event={"ID":"da92e42f-68d5-4f34-bcf5-c5a8e765e383","Type":"ContainerStarted","Data":"55d6b34ade76b9e8685e052d887d3efd3ffcd54d5c1e658fbf13cd48af4e6f92"} Dec 05 11:02:33.616093 master-0 kubenswrapper[24928]: I1205 11:02:33.615358 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" event={"ID":"68a107e3-b0eb-43fd-844f-817cf01ef767","Type":"ContainerStarted","Data":"db6733a832500ce17765e2107d717d32d7e105db2082025adc6c823935f68f8c"} Dec 05 11:02:33.631431 master-0 kubenswrapper[24928]: I1205 11:02:33.631377 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb" event={"ID":"5af60df4-8ca0-4bfb-8302-d3f9dcd4ef7a","Type":"ContainerStarted","Data":"ba79b7059ea1d2aa72014decfba905352b194adc48abcf7a05ba675f4ac6b967"} Dec 05 11:02:33.633569 master-0 kubenswrapper[24928]: I1205 11:02:33.633512 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" event={"ID":"fa9681c6-d40e-4075-9230-1ed7431a06fb","Type":"ContainerStarted","Data":"78c10e7068fe70e25b767527b274d5b5fa487a227f4564e76815ca059d792482"} Dec 05 11:02:34.406138 master-0 kubenswrapper[24928]: I1205 11:02:34.406005 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:02:34.406391 master-0 kubenswrapper[24928]: E1205 11:02:34.406270 24928 secret.go:189] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:34.406391 master-0 kubenswrapper[24928]: E1205 11:02:34.406347 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert podName:da186c71-5e15-45a3-ad76-e51aad43439f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:38.406320531 +0000 UTC m=+918.409514412 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert") pod "infra-operator-controller-manager-7d9c9d7fd8-4ht2g" (UID: "da186c71-5e15-45a3-ad76-e51aad43439f") : secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:34.913973 master-0 kubenswrapper[24928]: I1205 11:02:34.913904 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:02:34.914538 master-0 kubenswrapper[24928]: E1205 11:02:34.914106 24928 secret.go:189] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:34.914538 master-0 kubenswrapper[24928]: E1205 11:02:34.914181 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert podName:5b821d96-dc12-4794-a863-0cc47d4d8d65 nodeName:}" failed. No retries permitted until 2025-12-05 11:02:38.914159867 +0000 UTC m=+918.917353718 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert") pod "openstack-baremetal-operator-controller-manager-6f998f574688x6w" (UID: "5b821d96-dc12-4794-a863-0cc47d4d8d65") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:34.947684 master-0 kubenswrapper[24928]: W1205 11:02:34.947612 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd637a6d_86b1_4854_84d5_a86c73fda74a.slice/crio-0a0c11f56f6f509ab8fdacaa961afb9458d9cff0862e62e973ece0d7862fa277 WatchSource:0}: Error finding container 0a0c11f56f6f509ab8fdacaa961afb9458d9cff0862e62e973ece0d7862fa277: Status 404 returned error can't find the container with id 0a0c11f56f6f509ab8fdacaa961afb9458d9cff0862e62e973ece0d7862fa277 Dec 05 11:02:34.960089 master-0 kubenswrapper[24928]: W1205 11:02:34.960019 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16b00867_62d3_469b_b9c1_98c2e8a6bf47.slice/crio-1d2ec1f64d5e3c1b3679c9f21dde9649bf660e0af4a9a14dd7b8cbceea20c992 WatchSource:0}: Error finding container 1d2ec1f64d5e3c1b3679c9f21dde9649bf660e0af4a9a14dd7b8cbceea20c992: Status 404 returned error can't find the container with id 1d2ec1f64d5e3c1b3679c9f21dde9649bf660e0af4a9a14dd7b8cbceea20c992 Dec 05 11:02:34.964014 master-0 kubenswrapper[24928]: W1205 11:02:34.963909 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacbb966c_9984_42a4_9739_1cf87d7db982.slice/crio-8a1575ca8cf97a0272feb8d1e8ee19eff774789665af5a6c4b57329d7f5e3421 WatchSource:0}: Error finding container 8a1575ca8cf97a0272feb8d1e8ee19eff774789665af5a6c4b57329d7f5e3421: Status 404 returned error can't find the container with id 8a1575ca8cf97a0272feb8d1e8ee19eff774789665af5a6c4b57329d7f5e3421 Dec 05 11:02:35.320766 master-0 kubenswrapper[24928]: I1205 11:02:35.320685 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:35.321209 master-0 kubenswrapper[24928]: I1205 11:02:35.320794 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:35.321209 master-0 kubenswrapper[24928]: E1205 11:02:35.320877 24928 secret.go:189] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 11:02:35.321209 master-0 kubenswrapper[24928]: E1205 11:02:35.321002 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:39.320980294 +0000 UTC m=+919.324174225 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "webhook-server-cert" not found Dec 05 11:02:35.321209 master-0 kubenswrapper[24928]: E1205 11:02:35.321009 24928 secret.go:189] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 11:02:35.321209 master-0 kubenswrapper[24928]: E1205 11:02:35.321106 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:39.321082506 +0000 UTC m=+919.324276407 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "metrics-server-cert" not found Dec 05 11:02:35.654531 master-0 kubenswrapper[24928]: I1205 11:02:35.654290 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" event={"ID":"acbb966c-9984-42a4-9739-1cf87d7db982","Type":"ContainerStarted","Data":"8a1575ca8cf97a0272feb8d1e8ee19eff774789665af5a6c4b57329d7f5e3421"} Dec 05 11:02:35.655979 master-0 kubenswrapper[24928]: I1205 11:02:35.655863 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-78955d896f-8fcxk" event={"ID":"5db6830c-28ff-4fe3-bf7e-1f417df8bfeb","Type":"ContainerStarted","Data":"2d10ed66713d3f7c647ce579bc14c2320f0ab9f74d4625f2a1afa876ba212119"} Dec 05 11:02:35.657205 master-0 kubenswrapper[24928]: I1205 11:02:35.657147 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-696b999796-bwcl8" event={"ID":"cd637a6d-86b1-4854-84d5-a86c73fda74a","Type":"ContainerStarted","Data":"0a0c11f56f6f509ab8fdacaa961afb9458d9cff0862e62e973ece0d7862fa277"} Dec 05 11:02:35.658469 master-0 kubenswrapper[24928]: I1205 11:02:35.658404 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4" event={"ID":"16b00867-62d3-469b-b9c1-98c2e8a6bf47","Type":"ContainerStarted","Data":"1d2ec1f64d5e3c1b3679c9f21dde9649bf660e0af4a9a14dd7b8cbceea20c992"} Dec 05 11:02:35.660471 master-0 kubenswrapper[24928]: I1205 11:02:35.660396 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" event={"ID":"7cc35a07-e403-442b-bcdf-836525eff1f7","Type":"ContainerStarted","Data":"8ae3a22fe856241535ba689406c9c96b7652297635c5f0c4a3808bb0c70201dc"} Dec 05 11:02:38.421434 master-0 kubenswrapper[24928]: I1205 11:02:38.421354 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:02:38.422056 master-0 kubenswrapper[24928]: E1205 11:02:38.421541 24928 secret.go:189] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:38.422056 master-0 kubenswrapper[24928]: E1205 11:02:38.421596 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert podName:da186c71-5e15-45a3-ad76-e51aad43439f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:46.421579025 +0000 UTC m=+926.424772876 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert") pod "infra-operator-controller-manager-7d9c9d7fd8-4ht2g" (UID: "da186c71-5e15-45a3-ad76-e51aad43439f") : secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:38.931175 master-0 kubenswrapper[24928]: I1205 11:02:38.931067 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:02:38.931492 master-0 kubenswrapper[24928]: E1205 11:02:38.931410 24928 secret.go:189] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:38.931559 master-0 kubenswrapper[24928]: E1205 11:02:38.931503 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert podName:5b821d96-dc12-4794-a863-0cc47d4d8d65 nodeName:}" failed. No retries permitted until 2025-12-05 11:02:46.931478322 +0000 UTC m=+926.934672173 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert") pod "openstack-baremetal-operator-controller-manager-6f998f574688x6w" (UID: "5b821d96-dc12-4794-a863-0cc47d4d8d65") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:39.341378 master-0 kubenswrapper[24928]: I1205 11:02:39.341291 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:39.341657 master-0 kubenswrapper[24928]: I1205 11:02:39.341436 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:39.341657 master-0 kubenswrapper[24928]: E1205 11:02:39.341476 24928 secret.go:189] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 11:02:39.341657 master-0 kubenswrapper[24928]: E1205 11:02:39.341551 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:47.34153534 +0000 UTC m=+927.344729181 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "webhook-server-cert" not found Dec 05 11:02:39.341657 master-0 kubenswrapper[24928]: E1205 11:02:39.341630 24928 secret.go:189] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 11:02:39.341865 master-0 kubenswrapper[24928]: E1205 11:02:39.341744 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:02:47.341724784 +0000 UTC m=+927.344918635 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "metrics-server-cert" not found Dec 05 11:02:46.424999 master-0 kubenswrapper[24928]: I1205 11:02:46.424921 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:02:46.425559 master-0 kubenswrapper[24928]: E1205 11:02:46.425133 24928 secret.go:189] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:46.425559 master-0 kubenswrapper[24928]: E1205 11:02:46.425197 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert podName:da186c71-5e15-45a3-ad76-e51aad43439f nodeName:}" failed. No retries permitted until 2025-12-05 11:03:02.425182254 +0000 UTC m=+942.428376105 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert") pod "infra-operator-controller-manager-7d9c9d7fd8-4ht2g" (UID: "da186c71-5e15-45a3-ad76-e51aad43439f") : secret "infra-operator-webhook-server-cert" not found Dec 05 11:02:46.934386 master-0 kubenswrapper[24928]: I1205 11:02:46.934295 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:02:46.934717 master-0 kubenswrapper[24928]: E1205 11:02:46.934514 24928 secret.go:189] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:46.934717 master-0 kubenswrapper[24928]: E1205 11:02:46.934618 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert podName:5b821d96-dc12-4794-a863-0cc47d4d8d65 nodeName:}" failed. No retries permitted until 2025-12-05 11:03:02.934595969 +0000 UTC m=+942.937789830 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert") pod "openstack-baremetal-operator-controller-manager-6f998f574688x6w" (UID: "5b821d96-dc12-4794-a863-0cc47d4d8d65") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 11:02:47.343048 master-0 kubenswrapper[24928]: I1205 11:02:47.342959 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:47.343270 master-0 kubenswrapper[24928]: E1205 11:02:47.343162 24928 secret.go:189] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 11:02:47.343270 master-0 kubenswrapper[24928]: I1205 11:02:47.343244 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:02:47.343385 master-0 kubenswrapper[24928]: E1205 11:02:47.343309 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:03:03.343280732 +0000 UTC m=+943.346474603 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "metrics-server-cert" not found Dec 05 11:02:47.343541 master-0 kubenswrapper[24928]: E1205 11:02:47.343483 24928 secret.go:189] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 11:02:47.343605 master-0 kubenswrapper[24928]: E1205 11:02:47.343594 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs podName:b5f9281b-074b-46fa-8a52-dcdd102f5b8f nodeName:}" failed. No retries permitted until 2025-12-05 11:03:03.34357026 +0000 UTC m=+943.346764111 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs") pod "openstack-operator-controller-manager-599cfccd85-8d692" (UID: "b5f9281b-074b-46fa-8a52-dcdd102f5b8f") : secret "webhook-server-cert" not found Dec 05 11:02:54.896026 master-0 kubenswrapper[24928]: I1205 11:02:54.895956 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w" event={"ID":"44343d25-46da-40a9-b0be-94861421d403","Type":"ContainerStarted","Data":"b04155ef844276f728af9b3e354bac9ea694375bbd8ad8c7757e44462d5d48a8"} Dec 05 11:02:54.898696 master-0 kubenswrapper[24928]: I1205 11:02:54.898616 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z" event={"ID":"3d85e846-d44c-49aa-8128-02566193d999","Type":"ContainerStarted","Data":"21538ffe6c6fa464759bf586dec408a718f0a1bcda5a241949835049ab9d0ca4"} Dec 05 11:02:54.900809 master-0 kubenswrapper[24928]: I1205 11:02:54.900757 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc" event={"ID":"61b1765d-2b89-4d57-85c0-a7f9afe632fe","Type":"ContainerStarted","Data":"1c35968d50ee85b7157ed0d055d0e0178b242fcfa7113b77e5a8b3ea5da6d7c2"} Dec 05 11:02:54.902719 master-0 kubenswrapper[24928]: I1205 11:02:54.902673 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc" event={"ID":"216bdcf3-b763-4293-8ac4-3e7eeeae0b5d","Type":"ContainerStarted","Data":"9e16085e085dc43bd499d250e78b8a32cd756341c3ad45c949340224bc5906fb"} Dec 05 11:02:55.165984 master-0 kubenswrapper[24928]: E1205 11:02:55.165052 24928 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-srtkc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-865fc86d5b-z8jv6_openstack-operators(fa9681c6-d40e-4075-9230-1ed7431a06fb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:02:55.165984 master-0 kubenswrapper[24928]: E1205 11:02:55.165183 24928 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tgs2h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-647f96877-gcg9w_openstack-operators(68a107e3-b0eb-43fd-844f-817cf01ef767): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:02:55.165984 master-0 kubenswrapper[24928]: E1205 11:02:55.165576 24928 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4vmh9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-647d75769b-dft2w_openstack-operators(442a2392-c758-4713-baa3-3e119ade664c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:02:55.168656 master-0 kubenswrapper[24928]: E1205 11:02:55.166392 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" podUID="68a107e3-b0eb-43fd-844f-817cf01ef767" Dec 05 11:02:55.168656 master-0 kubenswrapper[24928]: E1205 11:02:55.166723 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" podUID="fa9681c6-d40e-4075-9230-1ed7431a06fb" Dec 05 11:02:55.168656 master-0 kubenswrapper[24928]: E1205 11:02:55.166767 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" podUID="442a2392-c758-4713-baa3-3e119ade664c" Dec 05 11:02:55.175002 master-0 kubenswrapper[24928]: E1205 11:02:55.174815 24928 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v26t2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-6b64f6f645-xf7hs_openstack-operators(a4e20703-5c0d-44a1-94c6-f5b97782ff5c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:02:55.175002 master-0 kubenswrapper[24928]: E1205 11:02:55.174967 24928 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jp5k9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-57dfcdd5b8-rth9m_openstack-operators(7cc35a07-e403-442b-bcdf-836525eff1f7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:02:55.179849 master-0 kubenswrapper[24928]: E1205 11:02:55.179797 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" podUID="7cc35a07-e403-442b-bcdf-836525eff1f7" Dec 05 11:02:55.180036 master-0 kubenswrapper[24928]: E1205 11:02:55.179878 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" podUID="a4e20703-5c0d-44a1-94c6-f5b97782ff5c" Dec 05 11:02:55.194831 master-0 kubenswrapper[24928]: E1205 11:02:55.194767 24928 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4pdcx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6b9b669fdb-tsk7b_openstack-operators(acbb966c-9984-42a4-9739-1cf87d7db982): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 11:02:55.196894 master-0 kubenswrapper[24928]: E1205 11:02:55.196022 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" podUID="acbb966c-9984-42a4-9739-1cf87d7db982" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:55.998789 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4" event={"ID":"ada63c1e-313b-465a-a556-32b452de65ab","Type":"ContainerStarted","Data":"d948abaf323421f8674ee76cef001b06661ff74e68035b8ac5dedb3cd4c06cfe"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.024546 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" event={"ID":"fa9681c6-d40e-4075-9230-1ed7431a06fb","Type":"ContainerStarted","Data":"0b2f1a284dd97489febffc619eb87efad1b925b3faa7ee4838e1da1f8661e467"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.025479 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: E1205 11:02:56.026741 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" podUID="fa9681c6-d40e-4075-9230-1ed7431a06fb" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.033676 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" event={"ID":"7cc35a07-e403-442b-bcdf-836525eff1f7","Type":"ContainerStarted","Data":"574c8c3ae5b071900b102d96a5593a3be2184c7c11b28574ec8c8c4a473f9578"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.033758 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: E1205 11:02:56.035207 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" podUID="7cc35a07-e403-442b-bcdf-836525eff1f7" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.035384 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-78955d896f-8fcxk" event={"ID":"5db6830c-28ff-4fe3-bf7e-1f417df8bfeb","Type":"ContainerStarted","Data":"c2418fd54a9f5c52dd162999056933fe2ee9402003a192b08757afcbd3d303c7"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.056794 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" event={"ID":"a4e20703-5c0d-44a1-94c6-f5b97782ff5c","Type":"ContainerStarted","Data":"fc68c6a6bbcce22af78aeea7c27f087ae7495977a6a8c78ec425ec46b038e7eb"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.057782 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: E1205 11:02:56.058781 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" podUID="a4e20703-5c0d-44a1-94c6-f5b97782ff5c" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.059776 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" event={"ID":"442a2392-c758-4713-baa3-3e119ade664c","Type":"ContainerStarted","Data":"0af8aabab974def3034eef6bf6dd6cd5c034f707c2156bcd7b3f6f1a56a61123"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.059853 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: E1205 11:02:56.060965 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" podUID="442a2392-c758-4713-baa3-3e119ade664c" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.062480 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls" event={"ID":"da92e42f-68d5-4f34-bcf5-c5a8e765e383","Type":"ContainerStarted","Data":"59031dbcd1bdd46f1531e3a1cbc9d37cfce9fae94ed4d4ccddb40e124298eb4d"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.064121 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" event={"ID":"68a107e3-b0eb-43fd-844f-817cf01ef767","Type":"ContainerStarted","Data":"ae0ce2c6aef730dd5edf71c7a39a353ae15ef4effeeb18c28c11bfb6cab4d6d7"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.064313 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.065494 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb" event={"ID":"5af60df4-8ca0-4bfb-8302-d3f9dcd4ef7a","Type":"ContainerStarted","Data":"4ba06c555bfa1797f8e870a9f9e3ee40ff2d05c44db569a4f9d4f9b974c1697e"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: E1205 11:02:56.065629 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" podUID="68a107e3-b0eb-43fd-844f-817cf01ef767" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.066975 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-696b999796-bwcl8" event={"ID":"cd637a6d-86b1-4854-84d5-a86c73fda74a","Type":"ContainerStarted","Data":"955998dd5db0eeee81a3f8e3494114a2da4e41f7afcea1125f7dd8286f3d624e"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.091691 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v" event={"ID":"4f712f48-af26-4222-a231-ac989f78d18f","Type":"ContainerStarted","Data":"a44451849ec004e76a6fa7c39d7e201493a7c3ee089e92c9efabc909407bfeab"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.094052 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4" event={"ID":"16b00867-62d3-469b-b9c1-98c2e8a6bf47","Type":"ContainerStarted","Data":"d3f42a92da7d7ccdf5c7676d7109b4054c0b741244b0d04cd4486c69ba9671d5"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.095374 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2" event={"ID":"5017646a-fb39-440c-afe5-a4d6cee68ed4","Type":"ContainerStarted","Data":"1bba151163020de1a308fee1c8762947f95869e843648d9ef1624d0d51ece473"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.100016 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c" event={"ID":"be90920f-159a-4fac-9746-70846d306f22","Type":"ContainerStarted","Data":"c43c327da3d0933bf4d8e9126741bd8c4bbbc044ac2a6d121408adfeb1f9f555"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.101976 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc" event={"ID":"1d175522-0c08-40c5-a309-f1820406ea2f","Type":"ContainerStarted","Data":"5db73b0652e64d6274b38c01f67040d44078c7f59d51de6a04db2ee73df4eb8b"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.103855 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" event={"ID":"acbb966c-9984-42a4-9739-1cf87d7db982","Type":"ContainerStarted","Data":"f037ac1b7788a51925983725091b306d746c3f85fc11f20087610c0706e0baa3"} Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: I1205 11:02:56.104108 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" Dec 05 11:02:56.917775 master-0 kubenswrapper[24928]: E1205 11:02:56.105666 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" podUID="acbb966c-9984-42a4-9739-1cf87d7db982" Dec 05 11:02:57.118194 master-0 kubenswrapper[24928]: I1205 11:02:57.118096 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-78955d896f-8fcxk" podStartSLOduration=7.412791878 podStartE2EDuration="26.118076666s" podCreationTimestamp="2025-12-05 11:02:31 +0000 UTC" firstStartedPulling="2025-12-05 11:02:34.955310804 +0000 UTC m=+914.958504655" lastFinishedPulling="2025-12-05 11:02:53.660595602 +0000 UTC m=+933.663789443" observedRunningTime="2025-12-05 11:02:57.099297957 +0000 UTC m=+937.102491808" watchObservedRunningTime="2025-12-05 11:02:57.118076666 +0000 UTC m=+937.121270517" Dec 05 11:02:57.131531 master-0 kubenswrapper[24928]: E1205 11:02:57.129568 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" podUID="7cc35a07-e403-442b-bcdf-836525eff1f7" Dec 05 11:02:57.131531 master-0 kubenswrapper[24928]: E1205 11:02:57.129677 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" podUID="acbb966c-9984-42a4-9739-1cf87d7db982" Dec 05 11:02:57.146868 master-0 kubenswrapper[24928]: E1205 11:02:57.146781 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" podUID="fa9681c6-d40e-4075-9230-1ed7431a06fb" Dec 05 11:02:57.146937 master-0 kubenswrapper[24928]: E1205 11:02:57.146909 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" podUID="a4e20703-5c0d-44a1-94c6-f5b97782ff5c" Dec 05 11:02:57.147010 master-0 kubenswrapper[24928]: E1205 11:02:57.146984 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" podUID="442a2392-c758-4713-baa3-3e119ade664c" Dec 05 11:02:57.147069 master-0 kubenswrapper[24928]: E1205 11:02:57.147049 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" podUID="68a107e3-b0eb-43fd-844f-817cf01ef767" Dec 05 11:03:00.183788 master-0 kubenswrapper[24928]: I1205 11:03:00.183702 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc" event={"ID":"216bdcf3-b763-4293-8ac4-3e7eeeae0b5d","Type":"ContainerStarted","Data":"ed9c677cb2a6e4d65ee9297a128561b63706b0bebc3496eb347b5227a2b97e1a"} Dec 05 11:03:00.186533 master-0 kubenswrapper[24928]: I1205 11:03:00.186460 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc" Dec 05 11:03:00.188932 master-0 kubenswrapper[24928]: I1205 11:03:00.188361 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc" Dec 05 11:03:00.205007 master-0 kubenswrapper[24928]: I1205 11:03:00.204794 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-696b999796-bwcl8" Dec 05 11:03:00.207231 master-0 kubenswrapper[24928]: I1205 11:03:00.207169 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-696b999796-bwcl8" Dec 05 11:03:00.227883 master-0 kubenswrapper[24928]: I1205 11:03:00.227846 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w" Dec 05 11:03:00.228051 master-0 kubenswrapper[24928]: I1205 11:03:00.228030 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w" event={"ID":"44343d25-46da-40a9-b0be-94861421d403","Type":"ContainerStarted","Data":"70d2607d8b8196f53acd81069026725cf2aa8de34bf263cf42b00ce4c003a3b5"} Dec 05 11:03:00.229078 master-0 kubenswrapper[24928]: I1205 11:03:00.229012 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w" Dec 05 11:03:00.240641 master-0 kubenswrapper[24928]: I1205 11:03:00.240590 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb" event={"ID":"5af60df4-8ca0-4bfb-8302-d3f9dcd4ef7a","Type":"ContainerStarted","Data":"e464fc8e82cff44fa861be637a01ba814f8b632b4f308323a3a16c68f8b97a68"} Dec 05 11:03:00.240962 master-0 kubenswrapper[24928]: I1205 11:03:00.240915 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb" Dec 05 11:03:00.241519 master-0 kubenswrapper[24928]: I1205 11:03:00.241383 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-78cd4f7769-xpcsc" podStartSLOduration=2.255480467 podStartE2EDuration="30.241347583s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:31.792311825 +0000 UTC m=+911.795505676" lastFinishedPulling="2025-12-05 11:02:59.778178941 +0000 UTC m=+939.781372792" observedRunningTime="2025-12-05 11:03:00.213960031 +0000 UTC m=+940.217153882" watchObservedRunningTime="2025-12-05 11:03:00.241347583 +0000 UTC m=+940.244541444" Dec 05 11:03:00.248081 master-0 kubenswrapper[24928]: I1205 11:03:00.247447 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb" Dec 05 11:03:00.258520 master-0 kubenswrapper[24928]: I1205 11:03:00.258396 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-696b999796-bwcl8" podStartSLOduration=5.380483727 podStartE2EDuration="30.258371968s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:34.955334744 +0000 UTC m=+914.958528595" lastFinishedPulling="2025-12-05 11:02:59.833222995 +0000 UTC m=+939.836416836" observedRunningTime="2025-12-05 11:03:00.245863227 +0000 UTC m=+940.249057088" watchObservedRunningTime="2025-12-05 11:03:00.258371968 +0000 UTC m=+940.261565819" Dec 05 11:03:00.349248 master-0 kubenswrapper[24928]: I1205 11:03:00.349157 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7cdd6b54fb-9wfjb" podStartSLOduration=3.193992293 podStartE2EDuration="30.349131791s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:32.735765995 +0000 UTC m=+912.738959846" lastFinishedPulling="2025-12-05 11:02:59.890905493 +0000 UTC m=+939.894099344" observedRunningTime="2025-12-05 11:03:00.305884823 +0000 UTC m=+940.309078674" watchObservedRunningTime="2025-12-05 11:03:00.349131791 +0000 UTC m=+940.352325652" Dec 05 11:03:00.364932 master-0 kubenswrapper[24928]: I1205 11:03:00.363831 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84bc9f68f5-t8l7w" podStartSLOduration=2.313874412 podStartE2EDuration="30.363813638s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:31.800538089 +0000 UTC m=+911.803731940" lastFinishedPulling="2025-12-05 11:02:59.850477315 +0000 UTC m=+939.853671166" observedRunningTime="2025-12-05 11:03:00.341195884 +0000 UTC m=+940.344389735" watchObservedRunningTime="2025-12-05 11:03:00.363813638 +0000 UTC m=+940.367007479" Dec 05 11:03:01.214444 master-0 kubenswrapper[24928]: I1205 11:03:01.212047 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" Dec 05 11:03:01.250041 master-0 kubenswrapper[24928]: I1205 11:03:01.249914 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" Dec 05 11:03:01.254956 master-0 kubenswrapper[24928]: I1205 11:03:01.254602 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2" event={"ID":"5017646a-fb39-440c-afe5-a4d6cee68ed4","Type":"ContainerStarted","Data":"cec010818ed907a75178bb12746c43257628cd1469efd5584d77fda92749c478"} Dec 05 11:03:01.256289 master-0 kubenswrapper[24928]: I1205 11:03:01.255029 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2" Dec 05 11:03:01.258670 master-0 kubenswrapper[24928]: I1205 11:03:01.258631 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c" event={"ID":"be90920f-159a-4fac-9746-70846d306f22","Type":"ContainerStarted","Data":"db6df06833a9f221dd85332bac95a3c6a0eba8a25dcd49586498cc2289a91dd0"} Dec 05 11:03:01.259518 master-0 kubenswrapper[24928]: I1205 11:03:01.259494 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c" Dec 05 11:03:01.260207 master-0 kubenswrapper[24928]: I1205 11:03:01.260186 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2" Dec 05 11:03:01.263126 master-0 kubenswrapper[24928]: I1205 11:03:01.263083 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c" Dec 05 11:03:01.263834 master-0 kubenswrapper[24928]: I1205 11:03:01.263771 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-696b999796-bwcl8" event={"ID":"cd637a6d-86b1-4854-84d5-a86c73fda74a","Type":"ContainerStarted","Data":"16e2f5528128f81d2a3e61e50a0cb106886504739cdcff482b725477c9f1bfa1"} Dec 05 11:03:01.278603 master-0 kubenswrapper[24928]: I1205 11:03:01.271102 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4" event={"ID":"ada63c1e-313b-465a-a556-32b452de65ab","Type":"ContainerStarted","Data":"eb8832758e3587c3777c138968c9508ae931ee12ea72ab682336046c738ba07d"} Dec 05 11:03:01.278603 master-0 kubenswrapper[24928]: I1205 11:03:01.273226 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4" Dec 05 11:03:01.278603 master-0 kubenswrapper[24928]: I1205 11:03:01.276197 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z" event={"ID":"3d85e846-d44c-49aa-8128-02566193d999","Type":"ContainerStarted","Data":"637c7f2f89dbdb4eb980660da77c43e799ac1c7484832e99134fa9f04b0c051a"} Dec 05 11:03:01.278603 master-0 kubenswrapper[24928]: I1205 11:03:01.276523 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z" Dec 05 11:03:01.285328 master-0 kubenswrapper[24928]: I1205 11:03:01.285019 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4" Dec 05 11:03:01.285328 master-0 kubenswrapper[24928]: I1205 11:03:01.285201 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z" Dec 05 11:03:01.286755 master-0 kubenswrapper[24928]: I1205 11:03:01.286693 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc" event={"ID":"61b1765d-2b89-4d57-85c0-a7f9afe632fe","Type":"ContainerStarted","Data":"15dec822d682b0394d6b3a010e413c0865fca09b81a4f43f48cdd3ea0a1695ef"} Dec 05 11:03:01.286949 master-0 kubenswrapper[24928]: I1205 11:03:01.286915 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc" Dec 05 11:03:01.290445 master-0 kubenswrapper[24928]: I1205 11:03:01.289845 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc" Dec 05 11:03:01.291681 master-0 kubenswrapper[24928]: I1205 11:03:01.291155 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v" event={"ID":"4f712f48-af26-4222-a231-ac989f78d18f","Type":"ContainerStarted","Data":"cd8b7d99578faa7b4d30f0b8a35437168ab63c9d4742b782b8a57b04b6fb4cc7"} Dec 05 11:03:01.291681 master-0 kubenswrapper[24928]: I1205 11:03:01.291468 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v" Dec 05 11:03:01.297317 master-0 kubenswrapper[24928]: I1205 11:03:01.297220 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v" Dec 05 11:03:01.307595 master-0 kubenswrapper[24928]: I1205 11:03:01.305032 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc" event={"ID":"1d175522-0c08-40c5-a309-f1820406ea2f","Type":"ContainerStarted","Data":"c19e332dc272d14c6033afe49fcd976ef5dce8727a15a31d98961633a5f041e4"} Dec 05 11:03:01.307936 master-0 kubenswrapper[24928]: I1205 11:03:01.307771 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc" Dec 05 11:03:01.322914 master-0 kubenswrapper[24928]: I1205 11:03:01.322769 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc" Dec 05 11:03:01.336035 master-0 kubenswrapper[24928]: I1205 11:03:01.332769 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-f6cc97788-5lr6c" podStartSLOduration=3.003489112 podStartE2EDuration="31.332664662s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:31.876976526 +0000 UTC m=+911.880170377" lastFinishedPulling="2025-12-05 11:03:00.206152076 +0000 UTC m=+940.209345927" observedRunningTime="2025-12-05 11:03:01.308007627 +0000 UTC m=+941.311201478" watchObservedRunningTime="2025-12-05 11:03:01.332664662 +0000 UTC m=+941.335858533" Dec 05 11:03:01.336035 master-0 kubenswrapper[24928]: I1205 11:03:01.334448 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4" event={"ID":"16b00867-62d3-469b-b9c1-98c2e8a6bf47","Type":"ContainerStarted","Data":"b0ac4e68f7f6f48ae129e948786ace95a1b26c6c37eafa4f194646480c4a6698"} Dec 05 11:03:01.336035 master-0 kubenswrapper[24928]: I1205 11:03:01.334968 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4" Dec 05 11:03:01.347142 master-0 kubenswrapper[24928]: I1205 11:03:01.341063 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4" Dec 05 11:03:01.347142 master-0 kubenswrapper[24928]: I1205 11:03:01.341185 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls" event={"ID":"da92e42f-68d5-4f34-bcf5-c5a8e765e383","Type":"ContainerStarted","Data":"358047d6883d5bdae907dcfc61203b41c41fd71c798f1693a5f2b1d3b1676580"} Dec 05 11:03:01.347142 master-0 kubenswrapper[24928]: I1205 11:03:01.341945 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls" Dec 05 11:03:01.347142 master-0 kubenswrapper[24928]: I1205 11:03:01.346308 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls" Dec 05 11:03:01.349343 master-0 kubenswrapper[24928]: I1205 11:03:01.349167 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-58b8dcc5fb-vv6s4" podStartSLOduration=3.103312752 podStartE2EDuration="31.349140773s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:32.305254638 +0000 UTC m=+912.308448489" lastFinishedPulling="2025-12-05 11:03:00.551082669 +0000 UTC m=+940.554276510" observedRunningTime="2025-12-05 11:03:01.332384586 +0000 UTC m=+941.335578447" watchObservedRunningTime="2025-12-05 11:03:01.349140773 +0000 UTC m=+941.352334624" Dec 05 11:03:01.364668 master-0 kubenswrapper[24928]: I1205 11:03:01.358651 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-5cd89994b5-ssmd2" podStartSLOduration=2.74158158 podStartE2EDuration="31.358622439s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:31.811610306 +0000 UTC m=+911.814804157" lastFinishedPulling="2025-12-05 11:03:00.428651165 +0000 UTC m=+940.431845016" observedRunningTime="2025-12-05 11:03:01.353931403 +0000 UTC m=+941.357125274" watchObservedRunningTime="2025-12-05 11:03:01.358622439 +0000 UTC m=+941.361816290" Dec 05 11:03:01.412528 master-0 kubenswrapper[24928]: I1205 11:03:01.412229 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" Dec 05 11:03:01.413747 master-0 kubenswrapper[24928]: I1205 11:03:01.412811 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-845b79dc4f-dc9ls" podStartSLOduration=3.44282717 podStartE2EDuration="31.41278506s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:32.749115758 +0000 UTC m=+912.752309609" lastFinishedPulling="2025-12-05 11:03:00.719073648 +0000 UTC m=+940.722267499" observedRunningTime="2025-12-05 11:03:01.388753921 +0000 UTC m=+941.391947772" watchObservedRunningTime="2025-12-05 11:03:01.41278506 +0000 UTC m=+941.415978911" Dec 05 11:03:01.438556 master-0 kubenswrapper[24928]: I1205 11:03:01.434175 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-7fd96594c7-5k6gc" podStartSLOduration=2.870411593 podStartE2EDuration="31.434155093s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:31.879646623 +0000 UTC m=+911.882840474" lastFinishedPulling="2025-12-05 11:03:00.443390123 +0000 UTC m=+940.446583974" observedRunningTime="2025-12-05 11:03:01.424539723 +0000 UTC m=+941.427733574" watchObservedRunningTime="2025-12-05 11:03:01.434155093 +0000 UTC m=+941.437348944" Dec 05 11:03:01.457360 master-0 kubenswrapper[24928]: I1205 11:03:01.456949 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-7c9bfd6967-bhx8z" podStartSLOduration=3.11566663 podStartE2EDuration="31.456924011s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:32.321225166 +0000 UTC m=+912.324419017" lastFinishedPulling="2025-12-05 11:03:00.662482547 +0000 UTC m=+940.665676398" observedRunningTime="2025-12-05 11:03:01.445039905 +0000 UTC m=+941.448233756" watchObservedRunningTime="2025-12-05 11:03:01.456924011 +0000 UTC m=+941.460117862" Dec 05 11:03:01.501262 master-0 kubenswrapper[24928]: I1205 11:03:01.501183 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-56f9fbf74b-pwlgc" podStartSLOduration=3.70622632 podStartE2EDuration="31.501157885s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:32.305233137 +0000 UTC m=+912.308426988" lastFinishedPulling="2025-12-05 11:03:00.100164702 +0000 UTC m=+940.103358553" observedRunningTime="2025-12-05 11:03:01.498554819 +0000 UTC m=+941.501748690" watchObservedRunningTime="2025-12-05 11:03:01.501157885 +0000 UTC m=+941.504351736" Dec 05 11:03:01.528108 master-0 kubenswrapper[24928]: I1205 11:03:01.528035 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" Dec 05 11:03:01.541280 master-0 kubenswrapper[24928]: I1205 11:03:01.541174 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-7b5867bfc7-7gjc4" podStartSLOduration=6.236417896 podStartE2EDuration="31.541158593s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:34.962114013 +0000 UTC m=+914.965307864" lastFinishedPulling="2025-12-05 11:03:00.26685471 +0000 UTC m=+940.270048561" observedRunningTime="2025-12-05 11:03:01.53384703 +0000 UTC m=+941.537040881" watchObservedRunningTime="2025-12-05 11:03:01.541158593 +0000 UTC m=+941.544352444" Dec 05 11:03:01.621949 master-0 kubenswrapper[24928]: I1205 11:03:01.621844 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-f8856dd79-7582v" podStartSLOduration=2.991014422 podStartE2EDuration="31.621827094s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:31.363763937 +0000 UTC m=+911.366957788" lastFinishedPulling="2025-12-05 11:02:59.994576619 +0000 UTC m=+939.997770460" observedRunningTime="2025-12-05 11:03:01.581707644 +0000 UTC m=+941.584901495" watchObservedRunningTime="2025-12-05 11:03:01.621827094 +0000 UTC m=+941.625020955" Dec 05 11:03:01.650558 master-0 kubenswrapper[24928]: I1205 11:03:01.650471 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" Dec 05 11:03:01.738085 master-0 kubenswrapper[24928]: I1205 11:03:01.738017 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" Dec 05 11:03:02.352558 master-0 kubenswrapper[24928]: I1205 11:03:02.352504 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" event={"ID":"a4e20703-5c0d-44a1-94c6-f5b97782ff5c","Type":"ContainerStarted","Data":"9306572a2b543962c44375c571e5105169c27d4b1895e16333f3026ee399e20b"} Dec 05 11:03:02.355272 master-0 kubenswrapper[24928]: I1205 11:03:02.355223 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" event={"ID":"7cc35a07-e403-442b-bcdf-836525eff1f7","Type":"ContainerStarted","Data":"b82288f1ea8028450bb86cadd17f6e75641197f50830d1057b822b548a2da28a"} Dec 05 11:03:02.357620 master-0 kubenswrapper[24928]: I1205 11:03:02.357563 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" event={"ID":"acbb966c-9984-42a4-9739-1cf87d7db982","Type":"ContainerStarted","Data":"1a2512da84b19e68ac02e245ad44e3793710195a10fb5982cfcc07c482af913e"} Dec 05 11:03:02.359646 master-0 kubenswrapper[24928]: I1205 11:03:02.359597 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" event={"ID":"442a2392-c758-4713-baa3-3e119ade664c","Type":"ContainerStarted","Data":"02e10cf91058b6e77454c3c51f67bd07ac692dfc9f32cb6000838821f6ff199c"} Dec 05 11:03:02.361643 master-0 kubenswrapper[24928]: I1205 11:03:02.361610 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" event={"ID":"68a107e3-b0eb-43fd-844f-817cf01ef767","Type":"ContainerStarted","Data":"29196bf9a147646bd3033f73f879b9f25b6f4b85660db9289641d25dc5a220ff"} Dec 05 11:03:02.364566 master-0 kubenswrapper[24928]: I1205 11:03:02.364520 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" event={"ID":"fa9681c6-d40e-4075-9230-1ed7431a06fb","Type":"ContainerStarted","Data":"014b330b842ae5471796d8e7641b29fef2692a8d6e9dd8cbdf6ebaaea9b1c2dc"} Dec 05 11:03:02.385599 master-0 kubenswrapper[24928]: I1205 11:03:02.385466 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-6b64f6f645-xf7hs" podStartSLOduration=11.600142294 podStartE2EDuration="32.3854464s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:32.729808407 +0000 UTC m=+912.733002258" lastFinishedPulling="2025-12-05 11:02:53.515112513 +0000 UTC m=+933.518306364" observedRunningTime="2025-12-05 11:03:02.376193569 +0000 UTC m=+942.379387420" watchObservedRunningTime="2025-12-05 11:03:02.3854464 +0000 UTC m=+942.388640251" Dec 05 11:03:02.454151 master-0 kubenswrapper[24928]: I1205 11:03:02.453964 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:03:02.457678 master-0 kubenswrapper[24928]: I1205 11:03:02.457619 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/da186c71-5e15-45a3-ad76-e51aad43439f-cert\") pod \"infra-operator-controller-manager-7d9c9d7fd8-4ht2g\" (UID: \"da186c71-5e15-45a3-ad76-e51aad43439f\") " pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:03:02.491150 master-0 kubenswrapper[24928]: I1205 11:03:02.486091 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-57dfcdd5b8-rth9m" podStartSLOduration=13.869787151 podStartE2EDuration="32.486062909s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:34.955094558 +0000 UTC m=+914.958288419" lastFinishedPulling="2025-12-05 11:02:53.571370326 +0000 UTC m=+933.574564177" observedRunningTime="2025-12-05 11:03:02.465478886 +0000 UTC m=+942.468672747" watchObservedRunningTime="2025-12-05 11:03:02.486062909 +0000 UTC m=+942.489256760" Dec 05 11:03:02.499074 master-0 kubenswrapper[24928]: I1205 11:03:02.498992 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-647f96877-gcg9w" podStartSLOduration=11.658553179 podStartE2EDuration="32.498972812s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:32.728570795 +0000 UTC m=+912.731764646" lastFinishedPulling="2025-12-05 11:02:53.568990418 +0000 UTC m=+933.572184279" observedRunningTime="2025-12-05 11:03:02.495025633 +0000 UTC m=+942.498219484" watchObservedRunningTime="2025-12-05 11:03:02.498972812 +0000 UTC m=+942.502166663" Dec 05 11:03:02.533609 master-0 kubenswrapper[24928]: I1205 11:03:02.533313 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:03:02.556175 master-0 kubenswrapper[24928]: I1205 11:03:02.552792 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-865fc86d5b-z8jv6" podStartSLOduration=11.671751289 podStartE2EDuration="32.552768673s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:32.747133219 +0000 UTC m=+912.750327080" lastFinishedPulling="2025-12-05 11:02:53.628150613 +0000 UTC m=+933.631344464" observedRunningTime="2025-12-05 11:03:02.546507457 +0000 UTC m=+942.549701308" watchObservedRunningTime="2025-12-05 11:03:02.552768673 +0000 UTC m=+942.555962524" Dec 05 11:03:02.648296 master-0 kubenswrapper[24928]: I1205 11:03:02.648080 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-647d75769b-dft2w" podStartSLOduration=11.898236067 podStartE2EDuration="32.64805653s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:32.312783355 +0000 UTC m=+912.315977216" lastFinishedPulling="2025-12-05 11:02:53.062603828 +0000 UTC m=+933.065797679" observedRunningTime="2025-12-05 11:03:02.630253935 +0000 UTC m=+942.633447796" watchObservedRunningTime="2025-12-05 11:03:02.64805653 +0000 UTC m=+942.651250381" Dec 05 11:03:02.669684 master-0 kubenswrapper[24928]: I1205 11:03:02.668492 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6b9b669fdb-tsk7b" podStartSLOduration=14.075609774 podStartE2EDuration="32.668475469s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:02:34.966897652 +0000 UTC m=+914.970091503" lastFinishedPulling="2025-12-05 11:02:53.559763347 +0000 UTC m=+933.562957198" observedRunningTime="2025-12-05 11:03:02.659129176 +0000 UTC m=+942.662323027" watchObservedRunningTime="2025-12-05 11:03:02.668475469 +0000 UTC m=+942.671669320" Dec 05 11:03:02.974329 master-0 kubenswrapper[24928]: I1205 11:03:02.974205 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:03:02.978331 master-0 kubenswrapper[24928]: I1205 11:03:02.978251 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b821d96-dc12-4794-a863-0cc47d4d8d65-cert\") pod \"openstack-baremetal-operator-controller-manager-6f998f574688x6w\" (UID: \"5b821d96-dc12-4794-a863-0cc47d4d8d65\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:03:03.075587 master-0 kubenswrapper[24928]: I1205 11:03:03.075521 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g"] Dec 05 11:03:03.225490 master-0 kubenswrapper[24928]: I1205 11:03:03.225399 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:03:03.378198 master-0 kubenswrapper[24928]: I1205 11:03:03.378101 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" event={"ID":"da186c71-5e15-45a3-ad76-e51aad43439f","Type":"ContainerStarted","Data":"93e7cd942028902f2c9252df6a2977199b34db66c47bab503cb357b10564fa66"} Dec 05 11:03:03.382798 master-0 kubenswrapper[24928]: I1205 11:03:03.382751 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:03:03.383152 master-0 kubenswrapper[24928]: I1205 11:03:03.383122 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:03:03.386376 master-0 kubenswrapper[24928]: I1205 11:03:03.386304 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-metrics-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:03:03.387253 master-0 kubenswrapper[24928]: I1205 11:03:03.387211 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b5f9281b-074b-46fa-8a52-dcdd102f5b8f-webhook-certs\") pod \"openstack-operator-controller-manager-599cfccd85-8d692\" (UID: \"b5f9281b-074b-46fa-8a52-dcdd102f5b8f\") " pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:03:03.572643 master-0 kubenswrapper[24928]: I1205 11:03:03.572557 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:03:03.656601 master-0 kubenswrapper[24928]: I1205 11:03:03.656445 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w"] Dec 05 11:03:04.016885 master-0 kubenswrapper[24928]: I1205 11:03:04.016716 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692"] Dec 05 11:03:04.030039 master-0 kubenswrapper[24928]: W1205 11:03:04.029752 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5f9281b_074b_46fa_8a52_dcdd102f5b8f.slice/crio-93d96d730b57ecee6d529a6cc4753abf73bd84193957342cd0bc9e0a2ba1c902 WatchSource:0}: Error finding container 93d96d730b57ecee6d529a6cc4753abf73bd84193957342cd0bc9e0a2ba1c902: Status 404 returned error can't find the container with id 93d96d730b57ecee6d529a6cc4753abf73bd84193957342cd0bc9e0a2ba1c902 Dec 05 11:03:04.398104 master-0 kubenswrapper[24928]: I1205 11:03:04.397501 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" event={"ID":"b5f9281b-074b-46fa-8a52-dcdd102f5b8f","Type":"ContainerStarted","Data":"e5bcc0380352879d1bc06f9d827f7355a7427b69b8d97c7807fda218c3973ed0"} Dec 05 11:03:04.398104 master-0 kubenswrapper[24928]: I1205 11:03:04.397571 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" event={"ID":"b5f9281b-074b-46fa-8a52-dcdd102f5b8f","Type":"ContainerStarted","Data":"93d96d730b57ecee6d529a6cc4753abf73bd84193957342cd0bc9e0a2ba1c902"} Dec 05 11:03:04.399078 master-0 kubenswrapper[24928]: I1205 11:03:04.398356 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:03:04.402542 master-0 kubenswrapper[24928]: I1205 11:03:04.402214 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" event={"ID":"5b821d96-dc12-4794-a863-0cc47d4d8d65","Type":"ContainerStarted","Data":"cb0c7ebe0f4db8ab4ad18e295c8d699d3972122ce5778515a35a66445f80c626"} Dec 05 11:03:04.448669 master-0 kubenswrapper[24928]: I1205 11:03:04.448576 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" podStartSLOduration=33.448556276 podStartE2EDuration="33.448556276s" podCreationTimestamp="2025-12-05 11:02:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:03:04.437183603 +0000 UTC m=+944.440377474" watchObservedRunningTime="2025-12-05 11:03:04.448556276 +0000 UTC m=+944.451750117" Dec 05 11:03:05.419321 master-0 kubenswrapper[24928]: I1205 11:03:05.419173 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" event={"ID":"da186c71-5e15-45a3-ad76-e51aad43439f","Type":"ContainerStarted","Data":"76bc11f70a044ad42cbecc495ceb72b17d878f6077c00c546b5cf4f1d9a3e40f"} Dec 05 11:03:05.419321 master-0 kubenswrapper[24928]: I1205 11:03:05.419250 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" event={"ID":"da186c71-5e15-45a3-ad76-e51aad43439f","Type":"ContainerStarted","Data":"3ebc01affab9014b32066f19a42a79bce050f86c6a65ae31bcf7b1a7927fa8cb"} Dec 05 11:03:05.419909 master-0 kubenswrapper[24928]: I1205 11:03:05.419776 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:03:05.451941 master-0 kubenswrapper[24928]: I1205 11:03:05.451816 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" podStartSLOduration=33.702830968 podStartE2EDuration="35.451779478s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:03:03.078857985 +0000 UTC m=+943.082051836" lastFinishedPulling="2025-12-05 11:03:04.827806495 +0000 UTC m=+944.831000346" observedRunningTime="2025-12-05 11:03:05.438528297 +0000 UTC m=+945.441722138" watchObservedRunningTime="2025-12-05 11:03:05.451779478 +0000 UTC m=+945.454973329" Dec 05 11:03:06.430443 master-0 kubenswrapper[24928]: I1205 11:03:06.430341 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" event={"ID":"5b821d96-dc12-4794-a863-0cc47d4d8d65","Type":"ContainerStarted","Data":"ffe4e0414fca15d3d02cd2a0df0c52d5969c0b80fa93c8570662a5ce4c5cbafe"} Dec 05 11:03:06.430443 master-0 kubenswrapper[24928]: I1205 11:03:06.430412 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" event={"ID":"5b821d96-dc12-4794-a863-0cc47d4d8d65","Type":"ContainerStarted","Data":"04f579a76333eaddcc91ddee0a42f5c1ecd6d9da2f904f682c8fa1228ee89dcf"} Dec 05 11:03:06.472953 master-0 kubenswrapper[24928]: I1205 11:03:06.472380 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" podStartSLOduration=34.204849967 podStartE2EDuration="36.472360452s" podCreationTimestamp="2025-12-05 11:02:30 +0000 UTC" firstStartedPulling="2025-12-05 11:03:03.666282125 +0000 UTC m=+943.669475976" lastFinishedPulling="2025-12-05 11:03:05.93379261 +0000 UTC m=+945.936986461" observedRunningTime="2025-12-05 11:03:06.465381448 +0000 UTC m=+946.468575299" watchObservedRunningTime="2025-12-05 11:03:06.472360452 +0000 UTC m=+946.475554303" Dec 05 11:03:07.440582 master-0 kubenswrapper[24928]: I1205 11:03:07.440504 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:03:12.547566 master-0 kubenswrapper[24928]: I1205 11:03:12.547045 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7d9c9d7fd8-4ht2g" Dec 05 11:03:13.233447 master-0 kubenswrapper[24928]: I1205 11:03:13.233336 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6f998f574688x6w" Dec 05 11:03:13.581860 master-0 kubenswrapper[24928]: I1205 11:03:13.581786 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-599cfccd85-8d692" Dec 05 11:03:52.902231 master-0 kubenswrapper[24928]: E1205 11:03:52.902151 24928 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/NetworkManager-dispatcher.service\": RecentStats: unable to find data in memory cache]" Dec 05 11:03:56.521462 master-0 kubenswrapper[24928]: I1205 11:03:56.518543 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z"] Dec 05 11:03:56.521462 master-0 kubenswrapper[24928]: I1205 11:03:56.521262 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" Dec 05 11:03:56.528592 master-0 kubenswrapper[24928]: I1205 11:03:56.525023 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 05 11:03:56.528592 master-0 kubenswrapper[24928]: I1205 11:03:56.525454 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 05 11:03:56.528592 master-0 kubenswrapper[24928]: I1205 11:03:56.525770 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 05 11:03:56.532507 master-0 kubenswrapper[24928]: I1205 11:03:56.531979 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z"] Dec 05 11:03:56.610261 master-0 kubenswrapper[24928]: I1205 11:03:56.610171 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75d7c5dbd7-cll7k"] Dec 05 11:03:56.622100 master-0 kubenswrapper[24928]: I1205 11:03:56.620799 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:03:56.624101 master-0 kubenswrapper[24928]: I1205 11:03:56.623517 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75d7c5dbd7-cll7k"] Dec 05 11:03:56.627479 master-0 kubenswrapper[24928]: I1205 11:03:56.624833 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 05 11:03:56.659391 master-0 kubenswrapper[24928]: I1205 11:03:56.659327 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c62nm\" (UniqueName: \"kubernetes.io/projected/58cd70f8-7ec9-48c4-a30e-c0576a310df1-kube-api-access-c62nm\") pod \"dnsmasq-dns-5dbfd7c4bf-4ts4z\" (UID: \"58cd70f8-7ec9-48c4-a30e-c0576a310df1\") " pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" Dec 05 11:03:56.659391 master-0 kubenswrapper[24928]: I1205 11:03:56.659386 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58cd70f8-7ec9-48c4-a30e-c0576a310df1-config\") pod \"dnsmasq-dns-5dbfd7c4bf-4ts4z\" (UID: \"58cd70f8-7ec9-48c4-a30e-c0576a310df1\") " pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" Dec 05 11:03:56.760976 master-0 kubenswrapper[24928]: I1205 11:03:56.760805 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-config\") pod \"dnsmasq-dns-75d7c5dbd7-cll7k\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:03:56.760976 master-0 kubenswrapper[24928]: I1205 11:03:56.760942 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-dns-svc\") pod \"dnsmasq-dns-75d7c5dbd7-cll7k\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:03:56.761313 master-0 kubenswrapper[24928]: I1205 11:03:56.760985 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-762qv\" (UniqueName: \"kubernetes.io/projected/1371ffa7-e320-4d17-bec7-15ba8d41484b-kube-api-access-762qv\") pod \"dnsmasq-dns-75d7c5dbd7-cll7k\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:03:56.761638 master-0 kubenswrapper[24928]: I1205 11:03:56.761598 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c62nm\" (UniqueName: \"kubernetes.io/projected/58cd70f8-7ec9-48c4-a30e-c0576a310df1-kube-api-access-c62nm\") pod \"dnsmasq-dns-5dbfd7c4bf-4ts4z\" (UID: \"58cd70f8-7ec9-48c4-a30e-c0576a310df1\") " pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" Dec 05 11:03:56.761718 master-0 kubenswrapper[24928]: I1205 11:03:56.761640 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58cd70f8-7ec9-48c4-a30e-c0576a310df1-config\") pod \"dnsmasq-dns-5dbfd7c4bf-4ts4z\" (UID: \"58cd70f8-7ec9-48c4-a30e-c0576a310df1\") " pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" Dec 05 11:03:56.762823 master-0 kubenswrapper[24928]: I1205 11:03:56.762775 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58cd70f8-7ec9-48c4-a30e-c0576a310df1-config\") pod \"dnsmasq-dns-5dbfd7c4bf-4ts4z\" (UID: \"58cd70f8-7ec9-48c4-a30e-c0576a310df1\") " pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" Dec 05 11:03:56.777630 master-0 kubenswrapper[24928]: I1205 11:03:56.777526 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c62nm\" (UniqueName: \"kubernetes.io/projected/58cd70f8-7ec9-48c4-a30e-c0576a310df1-kube-api-access-c62nm\") pod \"dnsmasq-dns-5dbfd7c4bf-4ts4z\" (UID: \"58cd70f8-7ec9-48c4-a30e-c0576a310df1\") " pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" Dec 05 11:03:56.863196 master-0 kubenswrapper[24928]: I1205 11:03:56.863110 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-config\") pod \"dnsmasq-dns-75d7c5dbd7-cll7k\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:03:56.863196 master-0 kubenswrapper[24928]: I1205 11:03:56.863173 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-dns-svc\") pod \"dnsmasq-dns-75d7c5dbd7-cll7k\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:03:56.863468 master-0 kubenswrapper[24928]: I1205 11:03:56.863206 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-762qv\" (UniqueName: \"kubernetes.io/projected/1371ffa7-e320-4d17-bec7-15ba8d41484b-kube-api-access-762qv\") pod \"dnsmasq-dns-75d7c5dbd7-cll7k\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:03:56.864461 master-0 kubenswrapper[24928]: I1205 11:03:56.864439 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-config\") pod \"dnsmasq-dns-75d7c5dbd7-cll7k\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:03:56.864580 master-0 kubenswrapper[24928]: I1205 11:03:56.864516 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-dns-svc\") pod \"dnsmasq-dns-75d7c5dbd7-cll7k\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:03:56.889985 master-0 kubenswrapper[24928]: I1205 11:03:56.887348 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-762qv\" (UniqueName: \"kubernetes.io/projected/1371ffa7-e320-4d17-bec7-15ba8d41484b-kube-api-access-762qv\") pod \"dnsmasq-dns-75d7c5dbd7-cll7k\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:03:56.917719 master-0 kubenswrapper[24928]: I1205 11:03:56.917646 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" Dec 05 11:03:56.962186 master-0 kubenswrapper[24928]: I1205 11:03:56.959096 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:03:58.106825 master-0 kubenswrapper[24928]: I1205 11:03:58.104521 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z"] Dec 05 11:03:58.114968 master-0 kubenswrapper[24928]: I1205 11:03:58.114399 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75d7c5dbd7-cll7k"] Dec 05 11:03:59.064149 master-0 kubenswrapper[24928]: I1205 11:03:59.064063 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" event={"ID":"58cd70f8-7ec9-48c4-a30e-c0576a310df1","Type":"ContainerStarted","Data":"55870f545964f297d3a1bf42e6e02737e0642b6c2b137bd43f0e5862bcb48e53"} Dec 05 11:03:59.065657 master-0 kubenswrapper[24928]: I1205 11:03:59.065600 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" event={"ID":"1371ffa7-e320-4d17-bec7-15ba8d41484b","Type":"ContainerStarted","Data":"cddebbad6ad189bf3500676c83b7f276418b124f77eb752ffe99342d353a8693"} Dec 05 11:03:59.618485 master-0 kubenswrapper[24928]: I1205 11:03:59.614502 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z"] Dec 05 11:03:59.647806 master-0 kubenswrapper[24928]: I1205 11:03:59.647736 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f856ff86c-mztch"] Dec 05 11:03:59.651376 master-0 kubenswrapper[24928]: I1205 11:03:59.651289 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:03:59.676591 master-0 kubenswrapper[24928]: I1205 11:03:59.671829 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f856ff86c-mztch"] Dec 05 11:03:59.804445 master-0 kubenswrapper[24928]: I1205 11:03:59.804312 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpxmj\" (UniqueName: \"kubernetes.io/projected/02058033-98c5-41b9-8b62-a366ea30a271-kube-api-access-zpxmj\") pod \"dnsmasq-dns-6f856ff86c-mztch\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:03:59.820734 master-0 kubenswrapper[24928]: I1205 11:03:59.805057 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-dns-svc\") pod \"dnsmasq-dns-6f856ff86c-mztch\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:03:59.820734 master-0 kubenswrapper[24928]: I1205 11:03:59.805240 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-config\") pod \"dnsmasq-dns-6f856ff86c-mztch\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:03:59.916165 master-0 kubenswrapper[24928]: I1205 11:03:59.912049 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-config\") pod \"dnsmasq-dns-6f856ff86c-mztch\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:03:59.916165 master-0 kubenswrapper[24928]: I1205 11:03:59.912234 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpxmj\" (UniqueName: \"kubernetes.io/projected/02058033-98c5-41b9-8b62-a366ea30a271-kube-api-access-zpxmj\") pod \"dnsmasq-dns-6f856ff86c-mztch\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:03:59.916165 master-0 kubenswrapper[24928]: I1205 11:03:59.912403 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-dns-svc\") pod \"dnsmasq-dns-6f856ff86c-mztch\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:03:59.924487 master-0 kubenswrapper[24928]: I1205 11:03:59.923101 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-config\") pod \"dnsmasq-dns-6f856ff86c-mztch\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:03:59.924487 master-0 kubenswrapper[24928]: I1205 11:03:59.923512 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-dns-svc\") pod \"dnsmasq-dns-6f856ff86c-mztch\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:03:59.949934 master-0 kubenswrapper[24928]: I1205 11:03:59.949306 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpxmj\" (UniqueName: \"kubernetes.io/projected/02058033-98c5-41b9-8b62-a366ea30a271-kube-api-access-zpxmj\") pod \"dnsmasq-dns-6f856ff86c-mztch\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:04:00.006590 master-0 kubenswrapper[24928]: I1205 11:03:59.995791 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:04:00.019232 master-0 kubenswrapper[24928]: I1205 11:04:00.019173 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75d7c5dbd7-cll7k"] Dec 05 11:04:00.120978 master-0 kubenswrapper[24928]: I1205 11:04:00.120764 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-658bb5765c-62tp9"] Dec 05 11:04:00.168511 master-0 kubenswrapper[24928]: I1205 11:04:00.163655 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658bb5765c-62tp9"] Dec 05 11:04:00.168511 master-0 kubenswrapper[24928]: I1205 11:04:00.163867 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:00.321956 master-0 kubenswrapper[24928]: I1205 11:04:00.321794 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-dns-svc\") pod \"dnsmasq-dns-658bb5765c-62tp9\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:00.321956 master-0 kubenswrapper[24928]: I1205 11:04:00.321946 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-config\") pod \"dnsmasq-dns-658bb5765c-62tp9\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:00.321956 master-0 kubenswrapper[24928]: I1205 11:04:00.322015 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5ld5\" (UniqueName: \"kubernetes.io/projected/69a1f456-df8f-4ab9-927c-02110a0d74b2-kube-api-access-f5ld5\") pod \"dnsmasq-dns-658bb5765c-62tp9\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:00.424170 master-0 kubenswrapper[24928]: I1205 11:04:00.424045 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-dns-svc\") pod \"dnsmasq-dns-658bb5765c-62tp9\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:00.424170 master-0 kubenswrapper[24928]: I1205 11:04:00.424129 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-config\") pod \"dnsmasq-dns-658bb5765c-62tp9\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:00.424395 master-0 kubenswrapper[24928]: I1205 11:04:00.424177 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5ld5\" (UniqueName: \"kubernetes.io/projected/69a1f456-df8f-4ab9-927c-02110a0d74b2-kube-api-access-f5ld5\") pod \"dnsmasq-dns-658bb5765c-62tp9\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:00.425318 master-0 kubenswrapper[24928]: I1205 11:04:00.425285 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-config\") pod \"dnsmasq-dns-658bb5765c-62tp9\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:00.426514 master-0 kubenswrapper[24928]: I1205 11:04:00.426466 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-dns-svc\") pod \"dnsmasq-dns-658bb5765c-62tp9\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:00.445537 master-0 kubenswrapper[24928]: I1205 11:04:00.445466 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5ld5\" (UniqueName: \"kubernetes.io/projected/69a1f456-df8f-4ab9-927c-02110a0d74b2-kube-api-access-f5ld5\") pod \"dnsmasq-dns-658bb5765c-62tp9\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:00.560992 master-0 kubenswrapper[24928]: I1205 11:04:00.560920 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:00.763249 master-0 kubenswrapper[24928]: I1205 11:04:00.763194 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f856ff86c-mztch"] Dec 05 11:04:01.110765 master-0 kubenswrapper[24928]: I1205 11:04:01.110693 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" event={"ID":"02058033-98c5-41b9-8b62-a366ea30a271","Type":"ContainerStarted","Data":"a7c9b5b3df77acc1dc86ca41be49346bc3fb1e64a82e2e2dc19367f9e7550053"} Dec 05 11:04:01.218189 master-0 kubenswrapper[24928]: I1205 11:04:01.217980 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-658bb5765c-62tp9"] Dec 05 11:04:01.271452 master-0 kubenswrapper[24928]: W1205 11:04:01.254078 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69a1f456_df8f_4ab9_927c_02110a0d74b2.slice/crio-0799f1c2262cf5ad5111454b6bf7da7a6ecf3becdf466fdae055ba971dad1942 WatchSource:0}: Error finding container 0799f1c2262cf5ad5111454b6bf7da7a6ecf3becdf466fdae055ba971dad1942: Status 404 returned error can't find the container with id 0799f1c2262cf5ad5111454b6bf7da7a6ecf3becdf466fdae055ba971dad1942 Dec 05 11:04:02.134672 master-0 kubenswrapper[24928]: I1205 11:04:02.134603 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" event={"ID":"69a1f456-df8f-4ab9-927c-02110a0d74b2","Type":"ContainerStarted","Data":"0799f1c2262cf5ad5111454b6bf7da7a6ecf3becdf466fdae055ba971dad1942"} Dec 05 11:04:03.695118 master-0 kubenswrapper[24928]: I1205 11:04:03.694469 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 05 11:04:03.698722 master-0 kubenswrapper[24928]: I1205 11:04:03.698646 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 11:04:03.704872 master-0 kubenswrapper[24928]: I1205 11:04:03.704461 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 05 11:04:03.705640 master-0 kubenswrapper[24928]: I1205 11:04:03.705395 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 05 11:04:03.719325 master-0 kubenswrapper[24928]: I1205 11:04:03.718639 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 05 11:04:03.726624 master-0 kubenswrapper[24928]: I1205 11:04:03.726557 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a177c607-c2db-411d-b4ee-e714452fea56-config-data\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.727521 master-0 kubenswrapper[24928]: I1205 11:04:03.726823 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a177c607-c2db-411d-b4ee-e714452fea56-kolla-config\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.727521 master-0 kubenswrapper[24928]: I1205 11:04:03.726867 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a177c607-c2db-411d-b4ee-e714452fea56-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.727521 master-0 kubenswrapper[24928]: I1205 11:04:03.726906 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a177c607-c2db-411d-b4ee-e714452fea56-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.727896 master-0 kubenswrapper[24928]: I1205 11:04:03.727733 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n7bz\" (UniqueName: \"kubernetes.io/projected/a177c607-c2db-411d-b4ee-e714452fea56-kube-api-access-5n7bz\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.748038 master-0 kubenswrapper[24928]: I1205 11:04:03.743487 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 11:04:03.786230 master-0 kubenswrapper[24928]: I1205 11:04:03.786180 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 11:04:03.790685 master-0 kubenswrapper[24928]: I1205 11:04:03.790346 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 11:04:03.825310 master-0 kubenswrapper[24928]: I1205 11:04:03.816882 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 11:04:03.825310 master-0 kubenswrapper[24928]: I1205 11:04:03.821583 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 05 11:04:03.825310 master-0 kubenswrapper[24928]: I1205 11:04:03.822003 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 05 11:04:03.832322 master-0 kubenswrapper[24928]: I1205 11:04:03.831897 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 05 11:04:03.832322 master-0 kubenswrapper[24928]: I1205 11:04:03.832230 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 05 11:04:03.832658 master-0 kubenswrapper[24928]: I1205 11:04:03.832391 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 05 11:04:03.832658 master-0 kubenswrapper[24928]: I1205 11:04:03.832465 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 05 11:04:03.841249 master-0 kubenswrapper[24928]: I1205 11:04:03.840927 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n7bz\" (UniqueName: \"kubernetes.io/projected/a177c607-c2db-411d-b4ee-e714452fea56-kube-api-access-5n7bz\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.841249 master-0 kubenswrapper[24928]: I1205 11:04:03.841186 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a177c607-c2db-411d-b4ee-e714452fea56-config-data\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.841249 master-0 kubenswrapper[24928]: I1205 11:04:03.841237 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a177c607-c2db-411d-b4ee-e714452fea56-kolla-config\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.841491 master-0 kubenswrapper[24928]: I1205 11:04:03.841278 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a177c607-c2db-411d-b4ee-e714452fea56-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.841491 master-0 kubenswrapper[24928]: I1205 11:04:03.841348 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a177c607-c2db-411d-b4ee-e714452fea56-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.849229 master-0 kubenswrapper[24928]: I1205 11:04:03.849182 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a177c607-c2db-411d-b4ee-e714452fea56-config-data\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.850097 master-0 kubenswrapper[24928]: I1205 11:04:03.850038 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a177c607-c2db-411d-b4ee-e714452fea56-kolla-config\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.890616 master-0 kubenswrapper[24928]: I1205 11:04:03.888296 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n7bz\" (UniqueName: \"kubernetes.io/projected/a177c607-c2db-411d-b4ee-e714452fea56-kube-api-access-5n7bz\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.892443 master-0 kubenswrapper[24928]: I1205 11:04:03.888696 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/a177c607-c2db-411d-b4ee-e714452fea56-memcached-tls-certs\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.900230 master-0 kubenswrapper[24928]: I1205 11:04:03.899694 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a177c607-c2db-411d-b4ee-e714452fea56-combined-ca-bundle\") pod \"memcached-0\" (UID: \"a177c607-c2db-411d-b4ee-e714452fea56\") " pod="openstack/memcached-0" Dec 05 11:04:03.945232 master-0 kubenswrapper[24928]: I1205 11:04:03.945076 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5wmh\" (UniqueName: \"kubernetes.io/projected/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-kube-api-access-r5wmh\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:03.945232 master-0 kubenswrapper[24928]: I1205 11:04:03.945308 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:03.945685 master-0 kubenswrapper[24928]: I1205 11:04:03.945415 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:03.945685 master-0 kubenswrapper[24928]: I1205 11:04:03.945559 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:03.945685 master-0 kubenswrapper[24928]: I1205 11:04:03.945649 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-14320b8d-25eb-43c7-8c84-7eef1f1639d4\" (UniqueName: \"kubernetes.io/csi/topolvm.io^94d1cf7d-729d-4d2f-a5f1-271e96bc3e6b\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:03.945685 master-0 kubenswrapper[24928]: I1205 11:04:03.945670 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:03.945845 master-0 kubenswrapper[24928]: I1205 11:04:03.945750 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:03.945845 master-0 kubenswrapper[24928]: I1205 11:04:03.945769 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:03.945845 master-0 kubenswrapper[24928]: I1205 11:04:03.945794 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:03.945845 master-0 kubenswrapper[24928]: I1205 11:04:03.945815 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:03.947974 master-0 kubenswrapper[24928]: I1205 11:04:03.945866 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-config-data\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049208 master-0 kubenswrapper[24928]: I1205 11:04:04.048193 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049208 master-0 kubenswrapper[24928]: I1205 11:04:04.048302 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049208 master-0 kubenswrapper[24928]: I1205 11:04:04.048335 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-14320b8d-25eb-43c7-8c84-7eef1f1639d4\" (UniqueName: \"kubernetes.io/csi/topolvm.io^94d1cf7d-729d-4d2f-a5f1-271e96bc3e6b\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049208 master-0 kubenswrapper[24928]: I1205 11:04:04.048355 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049208 master-0 kubenswrapper[24928]: I1205 11:04:04.048386 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049208 master-0 kubenswrapper[24928]: I1205 11:04:04.048408 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049208 master-0 kubenswrapper[24928]: I1205 11:04:04.048451 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049208 master-0 kubenswrapper[24928]: I1205 11:04:04.048475 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049208 master-0 kubenswrapper[24928]: I1205 11:04:04.048501 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-config-data\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049208 master-0 kubenswrapper[24928]: I1205 11:04:04.048558 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5wmh\" (UniqueName: \"kubernetes.io/projected/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-kube-api-access-r5wmh\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049208 master-0 kubenswrapper[24928]: I1205 11:04:04.048584 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.049767 master-0 kubenswrapper[24928]: I1205 11:04:04.049333 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.051964 master-0 kubenswrapper[24928]: I1205 11:04:04.051450 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-config-data\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.052483 master-0 kubenswrapper[24928]: I1205 11:04:04.052302 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.052483 master-0 kubenswrapper[24928]: I1205 11:04:04.052327 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.055214 master-0 kubenswrapper[24928]: I1205 11:04:04.054670 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:04:04.055214 master-0 kubenswrapper[24928]: I1205 11:04:04.054724 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-14320b8d-25eb-43c7-8c84-7eef1f1639d4\" (UniqueName: \"kubernetes.io/csi/topolvm.io^94d1cf7d-729d-4d2f-a5f1-271e96bc3e6b\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/22f556d98e69e91bde824ffaccfe2c94708e8a13b2f77c225c428729df3a9559/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.055214 master-0 kubenswrapper[24928]: I1205 11:04:04.054774 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.066404 master-0 kubenswrapper[24928]: I1205 11:04:04.055386 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.066777 master-0 kubenswrapper[24928]: I1205 11:04:04.060873 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.066862 master-0 kubenswrapper[24928]: I1205 11:04:04.061474 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.066862 master-0 kubenswrapper[24928]: I1205 11:04:04.063727 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.085948 master-0 kubenswrapper[24928]: I1205 11:04:04.085744 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5wmh\" (UniqueName: \"kubernetes.io/projected/ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc-kube-api-access-r5wmh\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:04.086931 master-0 kubenswrapper[24928]: I1205 11:04:04.086530 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 11:04:05.212404 master-0 kubenswrapper[24928]: I1205 11:04:05.212265 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 11:04:05.215871 master-0 kubenswrapper[24928]: I1205 11:04:05.215795 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.229348 master-0 kubenswrapper[24928]: I1205 11:04:05.229272 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 05 11:04:05.229665 master-0 kubenswrapper[24928]: I1205 11:04:05.229451 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 05 11:04:05.229665 master-0 kubenswrapper[24928]: I1205 11:04:05.229542 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 05 11:04:05.229665 master-0 kubenswrapper[24928]: I1205 11:04:05.229597 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 05 11:04:05.230639 master-0 kubenswrapper[24928]: I1205 11:04:05.229806 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 05 11:04:05.233124 master-0 kubenswrapper[24928]: I1205 11:04:05.232866 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 05 11:04:05.236936 master-0 kubenswrapper[24928]: I1205 11:04:05.236870 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 11:04:05.282271 master-0 kubenswrapper[24928]: I1205 11:04:05.282213 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d39078ce-d99a-4e13-bbec-6b986c74a797-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.282649 master-0 kubenswrapper[24928]: I1205 11:04:05.282334 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.282649 master-0 kubenswrapper[24928]: I1205 11:04:05.282367 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d39078ce-d99a-4e13-bbec-6b986c74a797-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.282649 master-0 kubenswrapper[24928]: I1205 11:04:05.282393 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d39078ce-d99a-4e13-bbec-6b986c74a797-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.282649 master-0 kubenswrapper[24928]: I1205 11:04:05.282490 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj6vf\" (UniqueName: \"kubernetes.io/projected/d39078ce-d99a-4e13-bbec-6b986c74a797-kube-api-access-nj6vf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.283718 master-0 kubenswrapper[24928]: I1205 11:04:05.283670 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d84f1c93-8b3e-4d59-9731-a857204d3e13\" (UniqueName: \"kubernetes.io/csi/topolvm.io^d58b54d0-5e92-4595-96fb-24e30c6b0c00\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.283797 master-0 kubenswrapper[24928]: I1205 11:04:05.283754 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.283922 master-0 kubenswrapper[24928]: I1205 11:04:05.283864 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.283922 master-0 kubenswrapper[24928]: I1205 11:04:05.283899 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d39078ce-d99a-4e13-bbec-6b986c74a797-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.288327 master-0 kubenswrapper[24928]: I1205 11:04:05.283974 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.288327 master-0 kubenswrapper[24928]: I1205 11:04:05.284001 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d39078ce-d99a-4e13-bbec-6b986c74a797-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.386150 master-0 kubenswrapper[24928]: I1205 11:04:05.386076 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d39078ce-d99a-4e13-bbec-6b986c74a797-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.386373 master-0 kubenswrapper[24928]: I1205 11:04:05.386194 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d39078ce-d99a-4e13-bbec-6b986c74a797-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.386539 master-0 kubenswrapper[24928]: I1205 11:04:05.386508 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.386716 master-0 kubenswrapper[24928]: I1205 11:04:05.386603 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d39078ce-d99a-4e13-bbec-6b986c74a797-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.386716 master-0 kubenswrapper[24928]: I1205 11:04:05.386696 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d39078ce-d99a-4e13-bbec-6b986c74a797-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.386920 master-0 kubenswrapper[24928]: I1205 11:04:05.386771 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj6vf\" (UniqueName: \"kubernetes.io/projected/d39078ce-d99a-4e13-bbec-6b986c74a797-kube-api-access-nj6vf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.386920 master-0 kubenswrapper[24928]: I1205 11:04:05.386839 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d84f1c93-8b3e-4d59-9731-a857204d3e13\" (UniqueName: \"kubernetes.io/csi/topolvm.io^d58b54d0-5e92-4595-96fb-24e30c6b0c00\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.386920 master-0 kubenswrapper[24928]: I1205 11:04:05.386875 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.387066 master-0 kubenswrapper[24928]: I1205 11:04:05.386924 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.387066 master-0 kubenswrapper[24928]: I1205 11:04:05.386948 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d39078ce-d99a-4e13-bbec-6b986c74a797-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.387235 master-0 kubenswrapper[24928]: I1205 11:04:05.387170 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.387848 master-0 kubenswrapper[24928]: I1205 11:04:05.387803 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.388100 master-0 kubenswrapper[24928]: I1205 11:04:05.388056 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.389197 master-0 kubenswrapper[24928]: I1205 11:04:05.388883 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d39078ce-d99a-4e13-bbec-6b986c74a797-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.389628 master-0 kubenswrapper[24928]: I1205 11:04:05.389583 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:04:05.389706 master-0 kubenswrapper[24928]: I1205 11:04:05.389634 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d84f1c93-8b3e-4d59-9731-a857204d3e13\" (UniqueName: \"kubernetes.io/csi/topolvm.io^d58b54d0-5e92-4595-96fb-24e30c6b0c00\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/4627a40ea609b0e4cd5ce5e3113e12124f54a26e4351d3d53c624b769335ff84/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.390095 master-0 kubenswrapper[24928]: I1205 11:04:05.390061 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d39078ce-d99a-4e13-bbec-6b986c74a797-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.390768 master-0 kubenswrapper[24928]: I1205 11:04:05.390734 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d39078ce-d99a-4e13-bbec-6b986c74a797-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.391010 master-0 kubenswrapper[24928]: I1205 11:04:05.390936 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d39078ce-d99a-4e13-bbec-6b986c74a797-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.393616 master-0 kubenswrapper[24928]: I1205 11:04:05.391892 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.396612 master-0 kubenswrapper[24928]: I1205 11:04:05.395985 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d39078ce-d99a-4e13-bbec-6b986c74a797-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.406468 master-0 kubenswrapper[24928]: I1205 11:04:05.406405 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d39078ce-d99a-4e13-bbec-6b986c74a797-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.441810 master-0 kubenswrapper[24928]: I1205 11:04:05.438678 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj6vf\" (UniqueName: \"kubernetes.io/projected/d39078ce-d99a-4e13-bbec-6b986c74a797-kube-api-access-nj6vf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:05.585899 master-0 kubenswrapper[24928]: I1205 11:04:05.585823 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-14320b8d-25eb-43c7-8c84-7eef1f1639d4\" (UniqueName: \"kubernetes.io/csi/topolvm.io^94d1cf7d-729d-4d2f-a5f1-271e96bc3e6b\") pod \"rabbitmq-server-0\" (UID: \"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc\") " pod="openstack/rabbitmq-server-0" Dec 05 11:04:05.644612 master-0 kubenswrapper[24928]: I1205 11:04:05.644406 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 11:04:06.258874 master-0 kubenswrapper[24928]: I1205 11:04:06.257438 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 05 11:04:06.260653 master-0 kubenswrapper[24928]: I1205 11:04:06.260603 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 11:04:06.263104 master-0 kubenswrapper[24928]: I1205 11:04:06.263032 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 05 11:04:06.263104 master-0 kubenswrapper[24928]: I1205 11:04:06.263062 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 05 11:04:06.265383 master-0 kubenswrapper[24928]: I1205 11:04:06.265296 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 11:04:06.267499 master-0 kubenswrapper[24928]: I1205 11:04:06.267471 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 05 11:04:06.322503 master-0 kubenswrapper[24928]: I1205 11:04:06.322411 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.322503 master-0 kubenswrapper[24928]: I1205 11:04:06.322485 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.323088 master-0 kubenswrapper[24928]: I1205 11:04:06.323033 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-kolla-config\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.323157 master-0 kubenswrapper[24928]: I1205 11:04:06.323130 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwvng\" (UniqueName: \"kubernetes.io/projected/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-kube-api-access-zwvng\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.323264 master-0 kubenswrapper[24928]: I1205 11:04:06.323224 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d4ea302b-0ce4-46de-8e5b-9197fc2061f3\" (UniqueName: \"kubernetes.io/csi/topolvm.io^307aac9e-9ab4-4d61-8419-680ca4966cbb\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.323307 master-0 kubenswrapper[24928]: I1205 11:04:06.323282 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.323480 master-0 kubenswrapper[24928]: I1205 11:04:06.323431 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.323750 master-0 kubenswrapper[24928]: I1205 11:04:06.323694 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-config-data-default\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.425758 master-0 kubenswrapper[24928]: I1205 11:04:06.425695 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-kolla-config\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.426002 master-0 kubenswrapper[24928]: I1205 11:04:06.425767 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwvng\" (UniqueName: \"kubernetes.io/projected/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-kube-api-access-zwvng\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.426002 master-0 kubenswrapper[24928]: I1205 11:04:06.425794 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d4ea302b-0ce4-46de-8e5b-9197fc2061f3\" (UniqueName: \"kubernetes.io/csi/topolvm.io^307aac9e-9ab4-4d61-8419-680ca4966cbb\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.426264 master-0 kubenswrapper[24928]: I1205 11:04:06.426217 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.426351 master-0 kubenswrapper[24928]: I1205 11:04:06.426326 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.426466 master-0 kubenswrapper[24928]: I1205 11:04:06.426443 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-config-data-default\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.426561 master-0 kubenswrapper[24928]: I1205 11:04:06.426542 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.427532 master-0 kubenswrapper[24928]: I1205 11:04:06.426573 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.429522 master-0 kubenswrapper[24928]: I1205 11:04:06.428313 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-config-data-default\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.429522 master-0 kubenswrapper[24928]: I1205 11:04:06.428395 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.429659 master-0 kubenswrapper[24928]: I1205 11:04:06.429323 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.429799 master-0 kubenswrapper[24928]: I1205 11:04:06.429753 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-kolla-config\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.431243 master-0 kubenswrapper[24928]: I1205 11:04:06.430889 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:04:06.431243 master-0 kubenswrapper[24928]: I1205 11:04:06.430926 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.431243 master-0 kubenswrapper[24928]: I1205 11:04:06.430982 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d4ea302b-0ce4-46de-8e5b-9197fc2061f3\" (UniqueName: \"kubernetes.io/csi/topolvm.io^307aac9e-9ab4-4d61-8419-680ca4966cbb\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/8a156cd3e220eb7b92303b68739802a880048c5b0fb5a654fc0324fcb341d5f5/globalmount\"" pod="openstack/openstack-galera-0" Dec 05 11:04:06.431600 master-0 kubenswrapper[24928]: I1205 11:04:06.431550 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.460314 master-0 kubenswrapper[24928]: I1205 11:04:06.459559 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwvng\" (UniqueName: \"kubernetes.io/projected/a3d90bd2-4fb8-4f48-abe2-5eb0915af549-kube-api-access-zwvng\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:06.928242 master-0 kubenswrapper[24928]: I1205 11:04:06.928177 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d84f1c93-8b3e-4d59-9731-a857204d3e13\" (UniqueName: \"kubernetes.io/csi/topolvm.io^d58b54d0-5e92-4595-96fb-24e30c6b0c00\") pod \"rabbitmq-cell1-server-0\" (UID: \"d39078ce-d99a-4e13-bbec-6b986c74a797\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:07.097233 master-0 kubenswrapper[24928]: I1205 11:04:07.097101 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:07.554867 master-0 kubenswrapper[24928]: I1205 11:04:07.554664 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 11:04:07.558326 master-0 kubenswrapper[24928]: I1205 11:04:07.558263 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.570453 master-0 kubenswrapper[24928]: I1205 11:04:07.562161 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 05 11:04:07.570453 master-0 kubenswrapper[24928]: I1205 11:04:07.562293 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 05 11:04:07.570453 master-0 kubenswrapper[24928]: I1205 11:04:07.562457 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 05 11:04:07.593901 master-0 kubenswrapper[24928]: I1205 11:04:07.579842 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 11:04:07.648703 master-0 kubenswrapper[24928]: I1205 11:04:07.648625 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.648922 master-0 kubenswrapper[24928]: I1205 11:04:07.648819 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.648922 master-0 kubenswrapper[24928]: I1205 11:04:07.648878 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.648922 master-0 kubenswrapper[24928]: I1205 11:04:07.648918 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-afafad88-6dfb-4a46-a6f0-5a52b25531ac\" (UniqueName: \"kubernetes.io/csi/topolvm.io^aa6971d0-2a18-497d-baea-2a7e093aba62\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.649039 master-0 kubenswrapper[24928]: I1205 11:04:07.648938 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.649039 master-0 kubenswrapper[24928]: I1205 11:04:07.648974 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.649104 master-0 kubenswrapper[24928]: I1205 11:04:07.649047 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xx5q\" (UniqueName: \"kubernetes.io/projected/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-kube-api-access-5xx5q\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.649374 master-0 kubenswrapper[24928]: I1205 11:04:07.649294 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.750761 master-0 kubenswrapper[24928]: I1205 11:04:07.750683 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.751139 master-0 kubenswrapper[24928]: I1205 11:04:07.751111 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.751313 master-0 kubenswrapper[24928]: I1205 11:04:07.751287 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.751493 master-0 kubenswrapper[24928]: I1205 11:04:07.751471 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.751672 master-0 kubenswrapper[24928]: I1205 11:04:07.751638 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.751782 master-0 kubenswrapper[24928]: I1205 11:04:07.751758 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-afafad88-6dfb-4a46-a6f0-5a52b25531ac\" (UniqueName: \"kubernetes.io/csi/topolvm.io^aa6971d0-2a18-497d-baea-2a7e093aba62\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.751950 master-0 kubenswrapper[24928]: I1205 11:04:07.751927 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.752738 master-0 kubenswrapper[24928]: I1205 11:04:07.752093 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xx5q\" (UniqueName: \"kubernetes.io/projected/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-kube-api-access-5xx5q\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.753030 master-0 kubenswrapper[24928]: I1205 11:04:07.751989 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.753103 master-0 kubenswrapper[24928]: I1205 11:04:07.752274 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.753199 master-0 kubenswrapper[24928]: I1205 11:04:07.753014 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.754221 master-0 kubenswrapper[24928]: I1205 11:04:07.754162 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.754303 master-0 kubenswrapper[24928]: I1205 11:04:07.754244 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:04:07.754348 master-0 kubenswrapper[24928]: I1205 11:04:07.754323 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-afafad88-6dfb-4a46-a6f0-5a52b25531ac\" (UniqueName: \"kubernetes.io/csi/topolvm.io^aa6971d0-2a18-497d-baea-2a7e093aba62\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/e97ee07b9d908cef63956e9812f9c529492e4027652bbea17752ef19ffaa40ba/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.755753 master-0 kubenswrapper[24928]: I1205 11:04:07.755696 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.762590 master-0 kubenswrapper[24928]: I1205 11:04:07.762489 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:07.769674 master-0 kubenswrapper[24928]: I1205 11:04:07.769628 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xx5q\" (UniqueName: \"kubernetes.io/projected/cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92-kube-api-access-5xx5q\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:08.005827 master-0 kubenswrapper[24928]: I1205 11:04:08.005674 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d4ea302b-0ce4-46de-8e5b-9197fc2061f3\" (UniqueName: \"kubernetes.io/csi/topolvm.io^307aac9e-9ab4-4d61-8419-680ca4966cbb\") pod \"openstack-galera-0\" (UID: \"a3d90bd2-4fb8-4f48-abe2-5eb0915af549\") " pod="openstack/openstack-galera-0" Dec 05 11:04:08.095313 master-0 kubenswrapper[24928]: I1205 11:04:08.095196 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 11:04:09.155868 master-0 kubenswrapper[24928]: I1205 11:04:09.155746 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-afafad88-6dfb-4a46-a6f0-5a52b25531ac\" (UniqueName: \"kubernetes.io/csi/topolvm.io^aa6971d0-2a18-497d-baea-2a7e093aba62\") pod \"openstack-cell1-galera-0\" (UID: \"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92\") " pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:09.283495 master-0 kubenswrapper[24928]: I1205 11:04:09.282545 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rz85b"] Dec 05 11:04:09.284855 master-0 kubenswrapper[24928]: I1205 11:04:09.284321 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.289458 master-0 kubenswrapper[24928]: I1205 11:04:09.286976 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 05 11:04:09.289458 master-0 kubenswrapper[24928]: I1205 11:04:09.287035 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 05 11:04:09.302212 master-0 kubenswrapper[24928]: I1205 11:04:09.302143 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rz85b"] Dec 05 11:04:09.391905 master-0 kubenswrapper[24928]: I1205 11:04:09.384915 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:09.398485 master-0 kubenswrapper[24928]: I1205 11:04:09.398378 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-rvm7b"] Dec 05 11:04:09.401790 master-0 kubenswrapper[24928]: I1205 11:04:09.401738 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.420379 master-0 kubenswrapper[24928]: I1205 11:04:09.420295 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rvm7b"] Dec 05 11:04:09.427938 master-0 kubenswrapper[24928]: I1205 11:04:09.427872 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/24ba60fc-8a32-4d40-973f-ccabca97d566-var-run-ovn\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.428051 master-0 kubenswrapper[24928]: I1205 11:04:09.427948 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/24ba60fc-8a32-4d40-973f-ccabca97d566-var-log-ovn\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.428051 master-0 kubenswrapper[24928]: I1205 11:04:09.427990 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ba60fc-8a32-4d40-973f-ccabca97d566-combined-ca-bundle\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.428051 master-0 kubenswrapper[24928]: I1205 11:04:09.428022 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/24ba60fc-8a32-4d40-973f-ccabca97d566-var-run\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.428282 master-0 kubenswrapper[24928]: I1205 11:04:09.428055 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ba60fc-8a32-4d40-973f-ccabca97d566-ovn-controller-tls-certs\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.428282 master-0 kubenswrapper[24928]: I1205 11:04:09.428099 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b592j\" (UniqueName: \"kubernetes.io/projected/24ba60fc-8a32-4d40-973f-ccabca97d566-kube-api-access-b592j\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.428282 master-0 kubenswrapper[24928]: I1205 11:04:09.428159 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24ba60fc-8a32-4d40-973f-ccabca97d566-scripts\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.530307 master-0 kubenswrapper[24928]: I1205 11:04:09.530222 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ba60fc-8a32-4d40-973f-ccabca97d566-combined-ca-bundle\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.530307 master-0 kubenswrapper[24928]: I1205 11:04:09.530285 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/24ba60fc-8a32-4d40-973f-ccabca97d566-var-run\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.530587 master-0 kubenswrapper[24928]: I1205 11:04:09.530319 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ba60fc-8a32-4d40-973f-ccabca97d566-ovn-controller-tls-certs\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.530587 master-0 kubenswrapper[24928]: I1205 11:04:09.530350 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-etc-ovs\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.530587 master-0 kubenswrapper[24928]: I1205 11:04:09.530380 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b592j\" (UniqueName: \"kubernetes.io/projected/24ba60fc-8a32-4d40-973f-ccabca97d566-kube-api-access-b592j\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.530587 master-0 kubenswrapper[24928]: I1205 11:04:09.530440 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40924985-cc8e-49f9-9215-8aca88c585ee-scripts\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.530587 master-0 kubenswrapper[24928]: I1205 11:04:09.530468 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24ba60fc-8a32-4d40-973f-ccabca97d566-scripts\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.531900 master-0 kubenswrapper[24928]: I1205 11:04:09.531833 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/24ba60fc-8a32-4d40-973f-ccabca97d566-var-run\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.532104 master-0 kubenswrapper[24928]: I1205 11:04:09.532060 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-var-log\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.532212 master-0 kubenswrapper[24928]: I1205 11:04:09.532171 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-var-lib\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.532553 master-0 kubenswrapper[24928]: I1205 11:04:09.532287 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jwx9\" (UniqueName: \"kubernetes.io/projected/40924985-cc8e-49f9-9215-8aca88c585ee-kube-api-access-9jwx9\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.532553 master-0 kubenswrapper[24928]: I1205 11:04:09.532338 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-var-run\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.532553 master-0 kubenswrapper[24928]: I1205 11:04:09.532491 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/24ba60fc-8a32-4d40-973f-ccabca97d566-var-run-ovn\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.532553 master-0 kubenswrapper[24928]: I1205 11:04:09.532536 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/24ba60fc-8a32-4d40-973f-ccabca97d566-var-log-ovn\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.532826 master-0 kubenswrapper[24928]: I1205 11:04:09.532786 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/24ba60fc-8a32-4d40-973f-ccabca97d566-var-log-ovn\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.535065 master-0 kubenswrapper[24928]: I1205 11:04:09.535009 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ba60fc-8a32-4d40-973f-ccabca97d566-combined-ca-bundle\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.535243 master-0 kubenswrapper[24928]: I1205 11:04:09.535212 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/24ba60fc-8a32-4d40-973f-ccabca97d566-var-run-ovn\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.544440 master-0 kubenswrapper[24928]: I1205 11:04:09.544369 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ba60fc-8a32-4d40-973f-ccabca97d566-ovn-controller-tls-certs\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.563317 master-0 kubenswrapper[24928]: I1205 11:04:09.563247 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/24ba60fc-8a32-4d40-973f-ccabca97d566-scripts\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.565909 master-0 kubenswrapper[24928]: I1205 11:04:09.565837 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b592j\" (UniqueName: \"kubernetes.io/projected/24ba60fc-8a32-4d40-973f-ccabca97d566-kube-api-access-b592j\") pod \"ovn-controller-rz85b\" (UID: \"24ba60fc-8a32-4d40-973f-ccabca97d566\") " pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.633041 master-0 kubenswrapper[24928]: I1205 11:04:09.632096 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rz85b" Dec 05 11:04:09.636531 master-0 kubenswrapper[24928]: I1205 11:04:09.634437 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40924985-cc8e-49f9-9215-8aca88c585ee-scripts\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.636531 master-0 kubenswrapper[24928]: I1205 11:04:09.634497 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-var-log\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.636531 master-0 kubenswrapper[24928]: I1205 11:04:09.634560 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-var-lib\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.636531 master-0 kubenswrapper[24928]: I1205 11:04:09.634631 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jwx9\" (UniqueName: \"kubernetes.io/projected/40924985-cc8e-49f9-9215-8aca88c585ee-kube-api-access-9jwx9\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.636531 master-0 kubenswrapper[24928]: I1205 11:04:09.634659 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-var-run\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.636531 master-0 kubenswrapper[24928]: I1205 11:04:09.634739 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-etc-ovs\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.636531 master-0 kubenswrapper[24928]: I1205 11:04:09.634900 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-var-run\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.636531 master-0 kubenswrapper[24928]: I1205 11:04:09.634935 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-etc-ovs\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.636531 master-0 kubenswrapper[24928]: I1205 11:04:09.634940 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-var-lib\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.637695 master-0 kubenswrapper[24928]: I1205 11:04:09.637633 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/40924985-cc8e-49f9-9215-8aca88c585ee-var-log\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.638628 master-0 kubenswrapper[24928]: I1205 11:04:09.638543 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40924985-cc8e-49f9-9215-8aca88c585ee-scripts\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.655187 master-0 kubenswrapper[24928]: I1205 11:04:09.655059 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jwx9\" (UniqueName: \"kubernetes.io/projected/40924985-cc8e-49f9-9215-8aca88c585ee-kube-api-access-9jwx9\") pod \"ovn-controller-ovs-rvm7b\" (UID: \"40924985-cc8e-49f9-9215-8aca88c585ee\") " pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:09.732991 master-0 kubenswrapper[24928]: I1205 11:04:09.732905 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:12.440956 master-0 kubenswrapper[24928]: I1205 11:04:12.440890 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 11:04:13.545834 master-0 kubenswrapper[24928]: I1205 11:04:13.545746 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 11:04:13.548187 master-0 kubenswrapper[24928]: I1205 11:04:13.548157 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:13.551035 master-0 kubenswrapper[24928]: I1205 11:04:13.550966 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 05 11:04:13.551125 master-0 kubenswrapper[24928]: I1205 11:04:13.550999 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 05 11:04:13.551387 master-0 kubenswrapper[24928]: I1205 11:04:13.551124 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 05 11:04:13.552046 master-0 kubenswrapper[24928]: I1205 11:04:13.552022 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 05 11:04:14.647811 master-0 kubenswrapper[24928]: I1205 11:04:14.647735 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 11:04:15.162001 master-0 kubenswrapper[24928]: I1205 11:04:15.160955 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75506b30-c0c5-4cfd-b151-3dbe75a05a77-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.162001 master-0 kubenswrapper[24928]: I1205 11:04:15.161075 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/75506b30-c0c5-4cfd-b151-3dbe75a05a77-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.162001 master-0 kubenswrapper[24928]: I1205 11:04:15.161113 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75506b30-c0c5-4cfd-b151-3dbe75a05a77-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.162001 master-0 kubenswrapper[24928]: I1205 11:04:15.161176 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h55p9\" (UniqueName: \"kubernetes.io/projected/75506b30-c0c5-4cfd-b151-3dbe75a05a77-kube-api-access-h55p9\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.162001 master-0 kubenswrapper[24928]: I1205 11:04:15.161210 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75506b30-c0c5-4cfd-b151-3dbe75a05a77-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.162001 master-0 kubenswrapper[24928]: I1205 11:04:15.161231 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/75506b30-c0c5-4cfd-b151-3dbe75a05a77-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.162001 master-0 kubenswrapper[24928]: I1205 11:04:15.161262 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-59066e22-bdc8-4bcd-b551-c2987d367d73\" (UniqueName: \"kubernetes.io/csi/topolvm.io^20e095a1-0dbb-4204-af59-68e6c32a0bb9\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.162001 master-0 kubenswrapper[24928]: I1205 11:04:15.161345 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75506b30-c0c5-4cfd-b151-3dbe75a05a77-config\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.266464 master-0 kubenswrapper[24928]: I1205 11:04:15.263551 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75506b30-c0c5-4cfd-b151-3dbe75a05a77-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.266464 master-0 kubenswrapper[24928]: I1205 11:04:15.263657 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/75506b30-c0c5-4cfd-b151-3dbe75a05a77-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.266464 master-0 kubenswrapper[24928]: I1205 11:04:15.263704 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75506b30-c0c5-4cfd-b151-3dbe75a05a77-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.266464 master-0 kubenswrapper[24928]: I1205 11:04:15.263805 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h55p9\" (UniqueName: \"kubernetes.io/projected/75506b30-c0c5-4cfd-b151-3dbe75a05a77-kube-api-access-h55p9\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.266464 master-0 kubenswrapper[24928]: I1205 11:04:15.263836 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75506b30-c0c5-4cfd-b151-3dbe75a05a77-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.266464 master-0 kubenswrapper[24928]: I1205 11:04:15.263852 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/75506b30-c0c5-4cfd-b151-3dbe75a05a77-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.266464 master-0 kubenswrapper[24928]: I1205 11:04:15.263925 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75506b30-c0c5-4cfd-b151-3dbe75a05a77-config\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.268827 master-0 kubenswrapper[24928]: I1205 11:04:15.268008 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/75506b30-c0c5-4cfd-b151-3dbe75a05a77-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.268827 master-0 kubenswrapper[24928]: I1205 11:04:15.268785 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75506b30-c0c5-4cfd-b151-3dbe75a05a77-config\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.269401 master-0 kubenswrapper[24928]: I1205 11:04:15.269344 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75506b30-c0c5-4cfd-b151-3dbe75a05a77-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.269745 master-0 kubenswrapper[24928]: I1205 11:04:15.269693 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/75506b30-c0c5-4cfd-b151-3dbe75a05a77-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.271452 master-0 kubenswrapper[24928]: I1205 11:04:15.271379 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75506b30-c0c5-4cfd-b151-3dbe75a05a77-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.272362 master-0 kubenswrapper[24928]: I1205 11:04:15.272261 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/75506b30-c0c5-4cfd-b151-3dbe75a05a77-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.344987 master-0 kubenswrapper[24928]: I1205 11:04:15.344936 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h55p9\" (UniqueName: \"kubernetes.io/projected/75506b30-c0c5-4cfd-b151-3dbe75a05a77-kube-api-access-h55p9\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.365781 master-0 kubenswrapper[24928]: I1205 11:04:15.365696 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-59066e22-bdc8-4bcd-b551-c2987d367d73\" (UniqueName: \"kubernetes.io/csi/topolvm.io^20e095a1-0dbb-4204-af59-68e6c32a0bb9\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:15.367988 master-0 kubenswrapper[24928]: I1205 11:04:15.367949 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:04:15.368077 master-0 kubenswrapper[24928]: I1205 11:04:15.368026 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-59066e22-bdc8-4bcd-b551-c2987d367d73\" (UniqueName: \"kubernetes.io/csi/topolvm.io^20e095a1-0dbb-4204-af59-68e6c32a0bb9\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/6fbb8ba5790215a0fd81eed086fd6d521481b033de3f222f83be3ded08bcfa41/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:17.064452 master-0 kubenswrapper[24928]: I1205 11:04:17.064383 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-59066e22-bdc8-4bcd-b551-c2987d367d73\" (UniqueName: \"kubernetes.io/csi/topolvm.io^20e095a1-0dbb-4204-af59-68e6c32a0bb9\") pod \"ovsdbserver-nb-0\" (UID: \"75506b30-c0c5-4cfd-b151-3dbe75a05a77\") " pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:17.284455 master-0 kubenswrapper[24928]: I1205 11:04:17.284341 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 11:04:17.292225 master-0 kubenswrapper[24928]: I1205 11:04:17.292158 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.296872 master-0 kubenswrapper[24928]: I1205 11:04:17.296791 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 05 11:04:17.297053 master-0 kubenswrapper[24928]: I1205 11:04:17.296895 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 11:04:17.297053 master-0 kubenswrapper[24928]: I1205 11:04:17.296924 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 05 11:04:17.299565 master-0 kubenswrapper[24928]: I1205 11:04:17.299511 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 05 11:04:17.368826 master-0 kubenswrapper[24928]: I1205 11:04:17.368748 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pjbjl"] Dec 05 11:04:17.373999 master-0 kubenswrapper[24928]: I1205 11:04:17.373933 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:17.394202 master-0 kubenswrapper[24928]: I1205 11:04:17.392013 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pjbjl"] Dec 05 11:04:17.404379 master-0 kubenswrapper[24928]: I1205 11:04:17.404320 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-utilities\") pod \"community-operators-pjbjl\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:17.404708 master-0 kubenswrapper[24928]: I1205 11:04:17.404506 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-catalog-content\") pod \"community-operators-pjbjl\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:17.404708 master-0 kubenswrapper[24928]: I1205 11:04:17.404601 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fzmv\" (UniqueName: \"kubernetes.io/projected/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-kube-api-access-4fzmv\") pod \"community-operators-pjbjl\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:17.468671 master-0 kubenswrapper[24928]: I1205 11:04:17.468605 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:17.506434 master-0 kubenswrapper[24928]: I1205 11:04:17.506340 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.506674 master-0 kubenswrapper[24928]: I1205 11:04:17.506409 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-utilities\") pod \"community-operators-pjbjl\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:17.506674 master-0 kubenswrapper[24928]: I1205 11:04:17.506540 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.506674 master-0 kubenswrapper[24928]: I1205 11:04:17.506575 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.506674 master-0 kubenswrapper[24928]: I1205 11:04:17.506618 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-catalog-content\") pod \"community-operators-pjbjl\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:17.506871 master-0 kubenswrapper[24928]: I1205 11:04:17.506742 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fzmv\" (UniqueName: \"kubernetes.io/projected/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-kube-api-access-4fzmv\") pod \"community-operators-pjbjl\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:17.506871 master-0 kubenswrapper[24928]: I1205 11:04:17.506821 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.507343 master-0 kubenswrapper[24928]: I1205 11:04:17.507299 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-utilities\") pod \"community-operators-pjbjl\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:17.507394 master-0 kubenswrapper[24928]: I1205 11:04:17.507317 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.507556 master-0 kubenswrapper[24928]: I1205 11:04:17.507365 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-catalog-content\") pod \"community-operators-pjbjl\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:17.507721 master-0 kubenswrapper[24928]: I1205 11:04:17.507689 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7rmj\" (UniqueName: \"kubernetes.io/projected/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-kube-api-access-w7rmj\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.507790 master-0 kubenswrapper[24928]: I1205 11:04:17.507757 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-config\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.507871 master-0 kubenswrapper[24928]: I1205 11:04:17.507851 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-272366f1-7cf9-4c98-8a2c-93113850d15f\" (UniqueName: \"kubernetes.io/csi/topolvm.io^445d61b3-6152-4425-8cd5-5a33ad89fd2b\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.522734 master-0 kubenswrapper[24928]: I1205 11:04:17.522695 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fzmv\" (UniqueName: \"kubernetes.io/projected/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-kube-api-access-4fzmv\") pod \"community-operators-pjbjl\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:17.610107 master-0 kubenswrapper[24928]: I1205 11:04:17.609939 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.610107 master-0 kubenswrapper[24928]: I1205 11:04:17.610009 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.610371 master-0 kubenswrapper[24928]: I1205 11:04:17.610330 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7rmj\" (UniqueName: \"kubernetes.io/projected/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-kube-api-access-w7rmj\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.610510 master-0 kubenswrapper[24928]: I1205 11:04:17.610433 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-config\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.610576 master-0 kubenswrapper[24928]: I1205 11:04:17.610545 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-272366f1-7cf9-4c98-8a2c-93113850d15f\" (UniqueName: \"kubernetes.io/csi/topolvm.io^445d61b3-6152-4425-8cd5-5a33ad89fd2b\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.610668 master-0 kubenswrapper[24928]: I1205 11:04:17.610633 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.610731 master-0 kubenswrapper[24928]: I1205 11:04:17.610635 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.610784 master-0 kubenswrapper[24928]: I1205 11:04:17.610725 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.610835 master-0 kubenswrapper[24928]: I1205 11:04:17.610778 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.611485 master-0 kubenswrapper[24928]: I1205 11:04:17.611445 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.612228 master-0 kubenswrapper[24928]: I1205 11:04:17.612180 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-config\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.614688 master-0 kubenswrapper[24928]: I1205 11:04:17.614657 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.614772 master-0 kubenswrapper[24928]: I1205 11:04:17.614714 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:04:17.614772 master-0 kubenswrapper[24928]: I1205 11:04:17.614742 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-272366f1-7cf9-4c98-8a2c-93113850d15f\" (UniqueName: \"kubernetes.io/csi/topolvm.io^445d61b3-6152-4425-8cd5-5a33ad89fd2b\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/3d39131845e172f015db255ad7278dcf773857fae9ee79d40e43786f41cb1b23/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.616748 master-0 kubenswrapper[24928]: I1205 11:04:17.616676 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.617409 master-0 kubenswrapper[24928]: I1205 11:04:17.617366 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.628027 master-0 kubenswrapper[24928]: I1205 11:04:17.627933 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7rmj\" (UniqueName: \"kubernetes.io/projected/51ecb4c8-aa75-4e4f-8529-4492445a9bb4-kube-api-access-w7rmj\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:17.767787 master-0 kubenswrapper[24928]: I1205 11:04:17.766648 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:18.419228 master-0 kubenswrapper[24928]: I1205 11:04:18.419165 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d39078ce-d99a-4e13-bbec-6b986c74a797","Type":"ContainerStarted","Data":"eaa1b2b7776356ca9af93f1229300059eb4c1bb4afe9b69f6d5fb06a5b1c1be2"} Dec 05 11:04:19.005752 master-0 kubenswrapper[24928]: I1205 11:04:19.005697 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 11:04:19.049029 master-0 kubenswrapper[24928]: W1205 11:04:19.048630 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd7fe9b2_e51f_4f2e_b1f1_ad8f92d7ba92.slice/crio-ca9a4ee730adac15c34019be0b4e379313d3b67fea872713bd37d97d91fbbf96 WatchSource:0}: Error finding container ca9a4ee730adac15c34019be0b4e379313d3b67fea872713bd37d97d91fbbf96: Status 404 returned error can't find the container with id ca9a4ee730adac15c34019be0b4e379313d3b67fea872713bd37d97d91fbbf96 Dec 05 11:04:19.086049 master-0 kubenswrapper[24928]: I1205 11:04:19.085996 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-272366f1-7cf9-4c98-8a2c-93113850d15f\" (UniqueName: \"kubernetes.io/csi/topolvm.io^445d61b3-6152-4425-8cd5-5a33ad89fd2b\") pod \"ovsdbserver-sb-0\" (UID: \"51ecb4c8-aa75-4e4f-8529-4492445a9bb4\") " pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:19.119479 master-0 kubenswrapper[24928]: I1205 11:04:19.119333 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:19.138850 master-0 kubenswrapper[24928]: I1205 11:04:19.138703 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 11:04:19.152114 master-0 kubenswrapper[24928]: I1205 11:04:19.151994 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 11:04:19.437511 master-0 kubenswrapper[24928]: I1205 11:04:19.437448 24928 generic.go:334] "Generic (PLEG): container finished" podID="58cd70f8-7ec9-48c4-a30e-c0576a310df1" containerID="c5fa8195f5b1bf7a6a2263c989136cef1dcda36902f517a5a275bb0a6379a6ee" exitCode=0 Dec 05 11:04:19.438114 master-0 kubenswrapper[24928]: I1205 11:04:19.437536 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" event={"ID":"58cd70f8-7ec9-48c4-a30e-c0576a310df1","Type":"ContainerDied","Data":"c5fa8195f5b1bf7a6a2263c989136cef1dcda36902f517a5a275bb0a6379a6ee"} Dec 05 11:04:19.439875 master-0 kubenswrapper[24928]: I1205 11:04:19.439835 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92","Type":"ContainerStarted","Data":"ca9a4ee730adac15c34019be0b4e379313d3b67fea872713bd37d97d91fbbf96"} Dec 05 11:04:19.442280 master-0 kubenswrapper[24928]: I1205 11:04:19.442242 24928 generic.go:334] "Generic (PLEG): container finished" podID="69a1f456-df8f-4ab9-927c-02110a0d74b2" containerID="55edfa24d1315fdd7e7862bc5f37f4cce06eb081ab98aa28e20f9d7e06851004" exitCode=0 Dec 05 11:04:19.442357 master-0 kubenswrapper[24928]: I1205 11:04:19.442302 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" event={"ID":"69a1f456-df8f-4ab9-927c-02110a0d74b2","Type":"ContainerDied","Data":"55edfa24d1315fdd7e7862bc5f37f4cce06eb081ab98aa28e20f9d7e06851004"} Dec 05 11:04:19.447010 master-0 kubenswrapper[24928]: I1205 11:04:19.446949 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc","Type":"ContainerStarted","Data":"ed3bc7bd7a649222b70710f05563b3658f0ad7b180c0307e7b816a521744652f"} Dec 05 11:04:19.449902 master-0 kubenswrapper[24928]: I1205 11:04:19.448589 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a3d90bd2-4fb8-4f48-abe2-5eb0915af549","Type":"ContainerStarted","Data":"997d1dabb1dcb3ff7b404b1944cce1a21329072cd94074dfe9c853bc49fcc54f"} Dec 05 11:04:19.453809 master-0 kubenswrapper[24928]: I1205 11:04:19.453767 24928 generic.go:334] "Generic (PLEG): container finished" podID="1371ffa7-e320-4d17-bec7-15ba8d41484b" containerID="903fbc419b27e68786e6c8236a99ddbac8f476e2252dea04e1c2464ffcbc1ce8" exitCode=0 Dec 05 11:04:19.456077 master-0 kubenswrapper[24928]: I1205 11:04:19.453815 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" event={"ID":"1371ffa7-e320-4d17-bec7-15ba8d41484b","Type":"ContainerDied","Data":"903fbc419b27e68786e6c8236a99ddbac8f476e2252dea04e1c2464ffcbc1ce8"} Dec 05 11:04:19.588749 master-0 kubenswrapper[24928]: I1205 11:04:19.587852 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 11:04:19.599592 master-0 kubenswrapper[24928]: I1205 11:04:19.598595 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rz85b"] Dec 05 11:04:19.866921 master-0 kubenswrapper[24928]: I1205 11:04:19.866871 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-rvm7b"] Dec 05 11:04:19.870997 master-0 kubenswrapper[24928]: W1205 11:04:19.870937 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40924985_cc8e_49f9_9215_8aca88c585ee.slice/crio-7a6a09cca0cc07e71302dd279682bed3496a8e00e017e7c1c1e3f19f9ed29f72 WatchSource:0}: Error finding container 7a6a09cca0cc07e71302dd279682bed3496a8e00e017e7c1c1e3f19f9ed29f72: Status 404 returned error can't find the container with id 7a6a09cca0cc07e71302dd279682bed3496a8e00e017e7c1c1e3f19f9ed29f72 Dec 05 11:04:20.106078 master-0 kubenswrapper[24928]: I1205 11:04:20.104070 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pjbjl"] Dec 05 11:04:20.236723 master-0 kubenswrapper[24928]: I1205 11:04:20.236659 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 11:04:20.471519 master-0 kubenswrapper[24928]: I1205 11:04:20.471225 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a177c607-c2db-411d-b4ee-e714452fea56","Type":"ContainerStarted","Data":"7aed4bc4402c0c7c26e92de10a343cfac0e2c062a817889c8c2b30e8fee5cacf"} Dec 05 11:04:20.473692 master-0 kubenswrapper[24928]: I1205 11:04:20.473099 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rvm7b" event={"ID":"40924985-cc8e-49f9-9215-8aca88c585ee","Type":"ContainerStarted","Data":"7a6a09cca0cc07e71302dd279682bed3496a8e00e017e7c1c1e3f19f9ed29f72"} Dec 05 11:04:20.478083 master-0 kubenswrapper[24928]: I1205 11:04:20.478011 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" event={"ID":"69a1f456-df8f-4ab9-927c-02110a0d74b2","Type":"ContainerStarted","Data":"b1075c17728b1e1aac22815dba6f661231e9ef95c47a4d69f70ff208cad8b81a"} Dec 05 11:04:20.478325 master-0 kubenswrapper[24928]: I1205 11:04:20.478175 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:20.485012 master-0 kubenswrapper[24928]: I1205 11:04:20.484944 24928 generic.go:334] "Generic (PLEG): container finished" podID="02058033-98c5-41b9-8b62-a366ea30a271" containerID="c336e09829a33dc28360bc763604848c019dde17a8f26a6a500676279632363c" exitCode=0 Dec 05 11:04:20.485240 master-0 kubenswrapper[24928]: I1205 11:04:20.485077 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" event={"ID":"02058033-98c5-41b9-8b62-a366ea30a271","Type":"ContainerDied","Data":"c336e09829a33dc28360bc763604848c019dde17a8f26a6a500676279632363c"} Dec 05 11:04:20.489977 master-0 kubenswrapper[24928]: I1205 11:04:20.489945 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rz85b" event={"ID":"24ba60fc-8a32-4d40-973f-ccabca97d566","Type":"ContainerStarted","Data":"dc39d9f1da43cc1685ac93b48747f0518f2b4932b62b0942e103049caecf913c"} Dec 05 11:04:20.505026 master-0 kubenswrapper[24928]: I1205 11:04:20.504935 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" podStartSLOduration=3.018169725 podStartE2EDuration="20.504910182s" podCreationTimestamp="2025-12-05 11:04:00 +0000 UTC" firstStartedPulling="2025-12-05 11:04:01.257883661 +0000 UTC m=+1001.261077502" lastFinishedPulling="2025-12-05 11:04:18.744624098 +0000 UTC m=+1018.747817959" observedRunningTime="2025-12-05 11:04:20.501549988 +0000 UTC m=+1020.504743869" watchObservedRunningTime="2025-12-05 11:04:20.504910182 +0000 UTC m=+1020.508104033" Dec 05 11:04:21.131938 master-0 kubenswrapper[24928]: I1205 11:04:21.131863 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 11:04:21.407760 master-0 kubenswrapper[24928]: I1205 11:04:21.407674 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-bjmv9"] Dec 05 11:04:21.414445 master-0 kubenswrapper[24928]: I1205 11:04:21.410614 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.415214 master-0 kubenswrapper[24928]: I1205 11:04:21.415103 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 05 11:04:21.421512 master-0 kubenswrapper[24928]: I1205 11:04:21.421039 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-bjmv9"] Dec 05 11:04:21.522551 master-0 kubenswrapper[24928]: I1205 11:04:21.522503 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" event={"ID":"58cd70f8-7ec9-48c4-a30e-c0576a310df1","Type":"ContainerDied","Data":"55870f545964f297d3a1bf42e6e02737e0642b6c2b137bd43f0e5862bcb48e53"} Dec 05 11:04:21.522551 master-0 kubenswrapper[24928]: I1205 11:04:21.522553 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55870f545964f297d3a1bf42e6e02737e0642b6c2b137bd43f0e5862bcb48e53" Dec 05 11:04:21.553566 master-0 kubenswrapper[24928]: I1205 11:04:21.549529 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f43649b-0385-4470-92ec-b031bb81f079-combined-ca-bundle\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.553566 master-0 kubenswrapper[24928]: I1205 11:04:21.549762 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8f43649b-0385-4470-92ec-b031bb81f079-ovs-rundir\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.553566 master-0 kubenswrapper[24928]: I1205 11:04:21.549850 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f43649b-0385-4470-92ec-b031bb81f079-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.553566 master-0 kubenswrapper[24928]: I1205 11:04:21.550684 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8f43649b-0385-4470-92ec-b031bb81f079-ovn-rundir\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.553566 master-0 kubenswrapper[24928]: I1205 11:04:21.550740 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f43649b-0385-4470-92ec-b031bb81f079-config\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.553566 master-0 kubenswrapper[24928]: I1205 11:04:21.550889 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5s7t\" (UniqueName: \"kubernetes.io/projected/8f43649b-0385-4470-92ec-b031bb81f079-kube-api-access-q5s7t\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.596609 master-0 kubenswrapper[24928]: I1205 11:04:21.596539 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658bb5765c-62tp9"] Dec 05 11:04:21.651204 master-0 kubenswrapper[24928]: I1205 11:04:21.651135 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-669479795c-v7gc5"] Dec 05 11:04:21.662538 master-0 kubenswrapper[24928]: I1205 11:04:21.653609 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.662538 master-0 kubenswrapper[24928]: I1205 11:04:21.661475 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.663628 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f43649b-0385-4470-92ec-b031bb81f079-combined-ca-bundle\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.663694 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8f43649b-0385-4470-92ec-b031bb81f079-ovs-rundir\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.663724 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f43649b-0385-4470-92ec-b031bb81f079-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.663746 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-ovsdbserver-nb\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.663777 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t27hk\" (UniqueName: \"kubernetes.io/projected/7fd3da2c-d0c3-439a-8867-de018998db88-kube-api-access-t27hk\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.663806 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8f43649b-0385-4470-92ec-b031bb81f079-ovn-rundir\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.663829 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f43649b-0385-4470-92ec-b031bb81f079-config\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.663865 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-dns-svc\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.663920 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5s7t\" (UniqueName: \"kubernetes.io/projected/8f43649b-0385-4470-92ec-b031bb81f079-kube-api-access-q5s7t\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.663938 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-config\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.674609 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8f43649b-0385-4470-92ec-b031bb81f079-ovn-rundir\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.675300 master-0 kubenswrapper[24928]: I1205 11:04:21.674717 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8f43649b-0385-4470-92ec-b031bb81f079-ovs-rundir\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.691962 master-0 kubenswrapper[24928]: I1205 11:04:21.685223 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f43649b-0385-4470-92ec-b031bb81f079-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.691962 master-0 kubenswrapper[24928]: I1205 11:04:21.685954 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f43649b-0385-4470-92ec-b031bb81f079-config\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.691962 master-0 kubenswrapper[24928]: I1205 11:04:21.689782 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f43649b-0385-4470-92ec-b031bb81f079-combined-ca-bundle\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.703874 master-0 kubenswrapper[24928]: I1205 11:04:21.703828 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-669479795c-v7gc5"] Dec 05 11:04:21.715632 master-0 kubenswrapper[24928]: I1205 11:04:21.715185 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5s7t\" (UniqueName: \"kubernetes.io/projected/8f43649b-0385-4470-92ec-b031bb81f079-kube-api-access-q5s7t\") pod \"ovn-controller-metrics-bjmv9\" (UID: \"8f43649b-0385-4470-92ec-b031bb81f079\") " pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.753465 master-0 kubenswrapper[24928]: I1205 11:04:21.751829 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-bjmv9" Dec 05 11:04:21.765984 master-0 kubenswrapper[24928]: I1205 11:04:21.765909 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-ovsdbserver-nb\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.766180 master-0 kubenswrapper[24928]: I1205 11:04:21.766010 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t27hk\" (UniqueName: \"kubernetes.io/projected/7fd3da2c-d0c3-439a-8867-de018998db88-kube-api-access-t27hk\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.766180 master-0 kubenswrapper[24928]: I1205 11:04:21.766065 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-dns-svc\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.766180 master-0 kubenswrapper[24928]: I1205 11:04:21.766131 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-config\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.767529 master-0 kubenswrapper[24928]: I1205 11:04:21.767498 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-config\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.768332 master-0 kubenswrapper[24928]: I1205 11:04:21.768301 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-ovsdbserver-nb\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.769445 master-0 kubenswrapper[24928]: I1205 11:04:21.769397 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-dns-svc\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.798218 master-0 kubenswrapper[24928]: I1205 11:04:21.797789 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t27hk\" (UniqueName: \"kubernetes.io/projected/7fd3da2c-d0c3-439a-8867-de018998db88-kube-api-access-t27hk\") pod \"dnsmasq-dns-669479795c-v7gc5\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:21.801584 master-0 kubenswrapper[24928]: I1205 11:04:21.801493 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" Dec 05 11:04:21.903452 master-0 kubenswrapper[24928]: I1205 11:04:21.902795 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f856ff86c-mztch"] Dec 05 11:04:21.908547 master-0 kubenswrapper[24928]: I1205 11:04:21.908502 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:04:21.966343 master-0 kubenswrapper[24928]: I1205 11:04:21.966210 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57545c5d5f-j2xrl"] Dec 05 11:04:21.971458 master-0 kubenswrapper[24928]: E1205 11:04:21.966873 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1371ffa7-e320-4d17-bec7-15ba8d41484b" containerName="init" Dec 05 11:04:21.971458 master-0 kubenswrapper[24928]: I1205 11:04:21.966895 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="1371ffa7-e320-4d17-bec7-15ba8d41484b" containerName="init" Dec 05 11:04:21.971458 master-0 kubenswrapper[24928]: E1205 11:04:21.966918 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58cd70f8-7ec9-48c4-a30e-c0576a310df1" containerName="init" Dec 05 11:04:21.971458 master-0 kubenswrapper[24928]: I1205 11:04:21.966924 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="58cd70f8-7ec9-48c4-a30e-c0576a310df1" containerName="init" Dec 05 11:04:21.971458 master-0 kubenswrapper[24928]: I1205 11:04:21.967145 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="1371ffa7-e320-4d17-bec7-15ba8d41484b" containerName="init" Dec 05 11:04:21.971458 master-0 kubenswrapper[24928]: I1205 11:04:21.967180 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="58cd70f8-7ec9-48c4-a30e-c0576a310df1" containerName="init" Dec 05 11:04:21.971458 master-0 kubenswrapper[24928]: I1205 11:04:21.968206 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:21.971458 master-0 kubenswrapper[24928]: I1205 11:04:21.969880 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c62nm\" (UniqueName: \"kubernetes.io/projected/58cd70f8-7ec9-48c4-a30e-c0576a310df1-kube-api-access-c62nm\") pod \"58cd70f8-7ec9-48c4-a30e-c0576a310df1\" (UID: \"58cd70f8-7ec9-48c4-a30e-c0576a310df1\") " Dec 05 11:04:21.971458 master-0 kubenswrapper[24928]: I1205 11:04:21.970207 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58cd70f8-7ec9-48c4-a30e-c0576a310df1-config\") pod \"58cd70f8-7ec9-48c4-a30e-c0576a310df1\" (UID: \"58cd70f8-7ec9-48c4-a30e-c0576a310df1\") " Dec 05 11:04:21.974226 master-0 kubenswrapper[24928]: I1205 11:04:21.973556 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 05 11:04:21.981841 master-0 kubenswrapper[24928]: I1205 11:04:21.980582 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57545c5d5f-j2xrl"] Dec 05 11:04:22.139813 master-0 kubenswrapper[24928]: I1205 11:04:22.139747 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-762qv\" (UniqueName: \"kubernetes.io/projected/1371ffa7-e320-4d17-bec7-15ba8d41484b-kube-api-access-762qv\") pod \"1371ffa7-e320-4d17-bec7-15ba8d41484b\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " Dec 05 11:04:22.140104 master-0 kubenswrapper[24928]: I1205 11:04:22.139910 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-dns-svc\") pod \"1371ffa7-e320-4d17-bec7-15ba8d41484b\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " Dec 05 11:04:22.140542 master-0 kubenswrapper[24928]: I1205 11:04:22.140393 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-config\") pod \"1371ffa7-e320-4d17-bec7-15ba8d41484b\" (UID: \"1371ffa7-e320-4d17-bec7-15ba8d41484b\") " Dec 05 11:04:22.141750 master-0 kubenswrapper[24928]: I1205 11:04:22.141660 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58cd70f8-7ec9-48c4-a30e-c0576a310df1-kube-api-access-c62nm" (OuterVolumeSpecName: "kube-api-access-c62nm") pod "58cd70f8-7ec9-48c4-a30e-c0576a310df1" (UID: "58cd70f8-7ec9-48c4-a30e-c0576a310df1"). InnerVolumeSpecName "kube-api-access-c62nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:22.152027 master-0 kubenswrapper[24928]: I1205 11:04:22.151950 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1371ffa7-e320-4d17-bec7-15ba8d41484b-kube-api-access-762qv" (OuterVolumeSpecName: "kube-api-access-762qv") pod "1371ffa7-e320-4d17-bec7-15ba8d41484b" (UID: "1371ffa7-e320-4d17-bec7-15ba8d41484b"). InnerVolumeSpecName "kube-api-access-762qv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:22.174081 master-0 kubenswrapper[24928]: I1205 11:04:22.173997 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-config\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.174763 master-0 kubenswrapper[24928]: I1205 11:04:22.174691 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-sb\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.174942 master-0 kubenswrapper[24928]: I1205 11:04:22.174915 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-dns-svc\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.175014 master-0 kubenswrapper[24928]: I1205 11:04:22.174984 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-nb\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.175121 master-0 kubenswrapper[24928]: I1205 11:04:22.175092 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2g8d\" (UniqueName: \"kubernetes.io/projected/5337585b-1757-4276-99fc-672cf04a1a89-kube-api-access-t2g8d\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.176151 master-0 kubenswrapper[24928]: I1205 11:04:22.176088 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:22.181911 master-0 kubenswrapper[24928]: I1205 11:04:22.181687 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-762qv\" (UniqueName: \"kubernetes.io/projected/1371ffa7-e320-4d17-bec7-15ba8d41484b-kube-api-access-762qv\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:22.181911 master-0 kubenswrapper[24928]: I1205 11:04:22.181792 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c62nm\" (UniqueName: \"kubernetes.io/projected/58cd70f8-7ec9-48c4-a30e-c0576a310df1-kube-api-access-c62nm\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:22.185942 master-0 kubenswrapper[24928]: I1205 11:04:22.185410 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58cd70f8-7ec9-48c4-a30e-c0576a310df1-config" (OuterVolumeSpecName: "config") pod "58cd70f8-7ec9-48c4-a30e-c0576a310df1" (UID: "58cd70f8-7ec9-48c4-a30e-c0576a310df1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:22.202291 master-0 kubenswrapper[24928]: I1205 11:04:22.202030 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1371ffa7-e320-4d17-bec7-15ba8d41484b" (UID: "1371ffa7-e320-4d17-bec7-15ba8d41484b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:22.207518 master-0 kubenswrapper[24928]: I1205 11:04:22.207397 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-config" (OuterVolumeSpecName: "config") pod "1371ffa7-e320-4d17-bec7-15ba8d41484b" (UID: "1371ffa7-e320-4d17-bec7-15ba8d41484b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:22.284292 master-0 kubenswrapper[24928]: I1205 11:04:22.284034 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-nb\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.284292 master-0 kubenswrapper[24928]: I1205 11:04:22.284125 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2g8d\" (UniqueName: \"kubernetes.io/projected/5337585b-1757-4276-99fc-672cf04a1a89-kube-api-access-t2g8d\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.284292 master-0 kubenswrapper[24928]: I1205 11:04:22.284189 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-config\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.284292 master-0 kubenswrapper[24928]: I1205 11:04:22.284287 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-sb\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.284851 master-0 kubenswrapper[24928]: I1205 11:04:22.284369 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-dns-svc\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.284851 master-0 kubenswrapper[24928]: I1205 11:04:22.284447 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/58cd70f8-7ec9-48c4-a30e-c0576a310df1-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:22.284851 master-0 kubenswrapper[24928]: I1205 11:04:22.284460 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:22.284851 master-0 kubenswrapper[24928]: I1205 11:04:22.284470 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1371ffa7-e320-4d17-bec7-15ba8d41484b-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:22.285597 master-0 kubenswrapper[24928]: I1205 11:04:22.285498 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-dns-svc\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.289618 master-0 kubenswrapper[24928]: I1205 11:04:22.289578 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-sb\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.290107 master-0 kubenswrapper[24928]: I1205 11:04:22.289185 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-nb\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.290536 master-0 kubenswrapper[24928]: I1205 11:04:22.290476 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-config\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.302031 master-0 kubenswrapper[24928]: I1205 11:04:22.301958 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2g8d\" (UniqueName: \"kubernetes.io/projected/5337585b-1757-4276-99fc-672cf04a1a89-kube-api-access-t2g8d\") pod \"dnsmasq-dns-57545c5d5f-j2xrl\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.493921 master-0 kubenswrapper[24928]: W1205 11:04:22.493839 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f43649b_0385_4470_92ec_b031bb81f079.slice/crio-35b8feceefa73428513158f5052afde4fac96cd900174afecaa388f4c255a731 WatchSource:0}: Error finding container 35b8feceefa73428513158f5052afde4fac96cd900174afecaa388f4c255a731: Status 404 returned error can't find the container with id 35b8feceefa73428513158f5052afde4fac96cd900174afecaa388f4c255a731 Dec 05 11:04:22.530289 master-0 kubenswrapper[24928]: I1205 11:04:22.529846 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-bjmv9"] Dec 05 11:04:22.546399 master-0 kubenswrapper[24928]: I1205 11:04:22.546329 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" event={"ID":"1371ffa7-e320-4d17-bec7-15ba8d41484b","Type":"ContainerDied","Data":"cddebbad6ad189bf3500676c83b7f276418b124f77eb752ffe99342d353a8693"} Dec 05 11:04:22.546692 master-0 kubenswrapper[24928]: I1205 11:04:22.546437 24928 scope.go:117] "RemoveContainer" containerID="903fbc419b27e68786e6c8236a99ddbac8f476e2252dea04e1c2464ffcbc1ce8" Dec 05 11:04:22.546892 master-0 kubenswrapper[24928]: I1205 11:04:22.546860 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75d7c5dbd7-cll7k" Dec 05 11:04:22.553619 master-0 kubenswrapper[24928]: I1205 11:04:22.553536 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-bjmv9" event={"ID":"8f43649b-0385-4470-92ec-b031bb81f079","Type":"ContainerStarted","Data":"35b8feceefa73428513158f5052afde4fac96cd900174afecaa388f4c255a731"} Dec 05 11:04:22.558819 master-0 kubenswrapper[24928]: I1205 11:04:22.558760 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"51ecb4c8-aa75-4e4f-8529-4492445a9bb4","Type":"ContainerStarted","Data":"300a25c08e9460d815654866d751136a03c9a333da10d9df3bcc5d7bc7121b31"} Dec 05 11:04:22.565330 master-0 kubenswrapper[24928]: I1205 11:04:22.565264 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" event={"ID":"02058033-98c5-41b9-8b62-a366ea30a271","Type":"ContainerStarted","Data":"bc019b4bff4412a64eba2e8ff6fcfccbb7824ee0c88b30e066d46504c8da6e47"} Dec 05 11:04:22.565693 master-0 kubenswrapper[24928]: I1205 11:04:22.565664 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:04:22.566552 master-0 kubenswrapper[24928]: I1205 11:04:22.565804 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" podUID="02058033-98c5-41b9-8b62-a366ea30a271" containerName="dnsmasq-dns" containerID="cri-o://bc019b4bff4412a64eba2e8ff6fcfccbb7824ee0c88b30e066d46504c8da6e47" gracePeriod=10 Dec 05 11:04:22.568258 master-0 kubenswrapper[24928]: I1205 11:04:22.568229 24928 generic.go:334] "Generic (PLEG): container finished" podID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" containerID="209ae81a830f495c81f3d5f4b06cb324a020f5e0955cc58e03abb9233bf4bb8c" exitCode=0 Dec 05 11:04:22.568360 master-0 kubenswrapper[24928]: I1205 11:04:22.568332 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjbjl" event={"ID":"6d62e2eb-0395-4f42-9371-1d6ea1df3a09","Type":"ContainerDied","Data":"209ae81a830f495c81f3d5f4b06cb324a020f5e0955cc58e03abb9233bf4bb8c"} Dec 05 11:04:22.568917 master-0 kubenswrapper[24928]: I1205 11:04:22.568366 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjbjl" event={"ID":"6d62e2eb-0395-4f42-9371-1d6ea1df3a09","Type":"ContainerStarted","Data":"e0ad5630d6f12b9e69ea2daa6c141ae44606e41732559ebf1a5fcea279067b14"} Dec 05 11:04:22.570866 master-0 kubenswrapper[24928]: I1205 11:04:22.570830 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z" Dec 05 11:04:22.571000 master-0 kubenswrapper[24928]: I1205 11:04:22.570903 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"75506b30-c0c5-4cfd-b151-3dbe75a05a77","Type":"ContainerStarted","Data":"239c196174f9f71ac0c20d005d47204dd92ed7e192db32507e1eeac625965e62"} Dec 05 11:04:22.574945 master-0 kubenswrapper[24928]: I1205 11:04:22.574888 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" podUID="69a1f456-df8f-4ab9-927c-02110a0d74b2" containerName="dnsmasq-dns" containerID="cri-o://b1075c17728b1e1aac22815dba6f661231e9ef95c47a4d69f70ff208cad8b81a" gracePeriod=10 Dec 05 11:04:22.596340 master-0 kubenswrapper[24928]: I1205 11:04:22.596248 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:22.899633 master-0 kubenswrapper[24928]: E1205 11:04:22.897278 24928 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02058033_98c5_41b9_8b62_a366ea30a271.slice/crio-bc019b4bff4412a64eba2e8ff6fcfccbb7824ee0c88b30e066d46504c8da6e47.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69a1f456_df8f_4ab9_927c_02110a0d74b2.slice/crio-b1075c17728b1e1aac22815dba6f661231e9ef95c47a4d69f70ff208cad8b81a.scope\": RecentStats: unable to find data in memory cache]" Dec 05 11:04:23.436110 master-0 kubenswrapper[24928]: I1205 11:04:23.435998 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57545c5d5f-j2xrl"] Dec 05 11:04:23.447162 master-0 kubenswrapper[24928]: I1205 11:04:23.447099 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-669479795c-v7gc5"] Dec 05 11:04:23.454674 master-0 kubenswrapper[24928]: I1205 11:04:23.454501 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" podStartSLOduration=6.359329098 podStartE2EDuration="24.454478939s" podCreationTimestamp="2025-12-05 11:03:59 +0000 UTC" firstStartedPulling="2025-12-05 11:04:00.755942742 +0000 UTC m=+1000.759136593" lastFinishedPulling="2025-12-05 11:04:18.851092583 +0000 UTC m=+1018.854286434" observedRunningTime="2025-12-05 11:04:23.430777378 +0000 UTC m=+1023.433971229" watchObservedRunningTime="2025-12-05 11:04:23.454478939 +0000 UTC m=+1023.457672790" Dec 05 11:04:23.506516 master-0 kubenswrapper[24928]: I1205 11:04:23.505184 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75d7c5dbd7-cll7k"] Dec 05 11:04:23.520761 master-0 kubenswrapper[24928]: I1205 11:04:23.520649 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75d7c5dbd7-cll7k"] Dec 05 11:04:23.585191 master-0 kubenswrapper[24928]: I1205 11:04:23.585061 24928 generic.go:334] "Generic (PLEG): container finished" podID="69a1f456-df8f-4ab9-927c-02110a0d74b2" containerID="b1075c17728b1e1aac22815dba6f661231e9ef95c47a4d69f70ff208cad8b81a" exitCode=0 Dec 05 11:04:23.585191 master-0 kubenswrapper[24928]: I1205 11:04:23.585125 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" event={"ID":"69a1f456-df8f-4ab9-927c-02110a0d74b2","Type":"ContainerDied","Data":"b1075c17728b1e1aac22815dba6f661231e9ef95c47a4d69f70ff208cad8b81a"} Dec 05 11:04:23.587355 master-0 kubenswrapper[24928]: I1205 11:04:23.587307 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d39078ce-d99a-4e13-bbec-6b986c74a797","Type":"ContainerStarted","Data":"caa0367285949fe4ea582abe038e4d12d2481440aca0e346aa8b61fe6ea33fa4"} Dec 05 11:04:23.589974 master-0 kubenswrapper[24928]: I1205 11:04:23.589770 24928 generic.go:334] "Generic (PLEG): container finished" podID="02058033-98c5-41b9-8b62-a366ea30a271" containerID="bc019b4bff4412a64eba2e8ff6fcfccbb7824ee0c88b30e066d46504c8da6e47" exitCode=0 Dec 05 11:04:23.589974 master-0 kubenswrapper[24928]: I1205 11:04:23.589820 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" event={"ID":"02058033-98c5-41b9-8b62-a366ea30a271","Type":"ContainerDied","Data":"bc019b4bff4412a64eba2e8ff6fcfccbb7824ee0c88b30e066d46504c8da6e47"} Dec 05 11:04:23.736158 master-0 kubenswrapper[24928]: I1205 11:04:23.736001 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z"] Dec 05 11:04:23.751649 master-0 kubenswrapper[24928]: I1205 11:04:23.751568 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dbfd7c4bf-4ts4z"] Dec 05 11:04:24.251849 master-0 kubenswrapper[24928]: I1205 11:04:24.251778 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1371ffa7-e320-4d17-bec7-15ba8d41484b" path="/var/lib/kubelet/pods/1371ffa7-e320-4d17-bec7-15ba8d41484b/volumes" Dec 05 11:04:24.269473 master-0 kubenswrapper[24928]: I1205 11:04:24.267097 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58cd70f8-7ec9-48c4-a30e-c0576a310df1" path="/var/lib/kubelet/pods/58cd70f8-7ec9-48c4-a30e-c0576a310df1/volumes" Dec 05 11:04:24.614180 master-0 kubenswrapper[24928]: I1205 11:04:24.614119 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc","Type":"ContainerStarted","Data":"4704001173b50a7a4e8491dc21bc67559657f4c7792b6c90ee601e27644cfe75"} Dec 05 11:04:26.329257 master-0 kubenswrapper[24928]: W1205 11:04:26.328823 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5337585b_1757_4276_99fc_672cf04a1a89.slice/crio-4658264ef1211933bb6d627e7e5009cc5963b2b2e42c7337055e1a15d5d5d501 WatchSource:0}: Error finding container 4658264ef1211933bb6d627e7e5009cc5963b2b2e42c7337055e1a15d5d5d501: Status 404 returned error can't find the container with id 4658264ef1211933bb6d627e7e5009cc5963b2b2e42c7337055e1a15d5d5d501 Dec 05 11:04:26.332829 master-0 kubenswrapper[24928]: W1205 11:04:26.332746 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fd3da2c_d0c3_439a_8867_de018998db88.slice/crio-fd7c85ec1f9f9cc7bc9b46a8ed40d3cbb895478edbf33bc2f5c4b8249e08286a WatchSource:0}: Error finding container fd7c85ec1f9f9cc7bc9b46a8ed40d3cbb895478edbf33bc2f5c4b8249e08286a: Status 404 returned error can't find the container with id fd7c85ec1f9f9cc7bc9b46a8ed40d3cbb895478edbf33bc2f5c4b8249e08286a Dec 05 11:04:26.643616 master-0 kubenswrapper[24928]: I1205 11:04:26.643232 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" event={"ID":"5337585b-1757-4276-99fc-672cf04a1a89","Type":"ContainerStarted","Data":"4658264ef1211933bb6d627e7e5009cc5963b2b2e42c7337055e1a15d5d5d501"} Dec 05 11:04:26.645622 master-0 kubenswrapper[24928]: I1205 11:04:26.645568 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-669479795c-v7gc5" event={"ID":"7fd3da2c-d0c3-439a-8867-de018998db88","Type":"ContainerStarted","Data":"fd7c85ec1f9f9cc7bc9b46a8ed40d3cbb895478edbf33bc2f5c4b8249e08286a"} Dec 05 11:04:26.648106 master-0 kubenswrapper[24928]: I1205 11:04:26.648069 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" event={"ID":"69a1f456-df8f-4ab9-927c-02110a0d74b2","Type":"ContainerDied","Data":"0799f1c2262cf5ad5111454b6bf7da7a6ecf3becdf466fdae055ba971dad1942"} Dec 05 11:04:26.648106 master-0 kubenswrapper[24928]: I1205 11:04:26.648097 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0799f1c2262cf5ad5111454b6bf7da7a6ecf3becdf466fdae055ba971dad1942" Dec 05 11:04:26.663574 master-0 kubenswrapper[24928]: I1205 11:04:26.663522 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:26.734715 master-0 kubenswrapper[24928]: I1205 11:04:26.715357 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-config\") pod \"69a1f456-df8f-4ab9-927c-02110a0d74b2\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " Dec 05 11:04:26.734715 master-0 kubenswrapper[24928]: I1205 11:04:26.715454 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5ld5\" (UniqueName: \"kubernetes.io/projected/69a1f456-df8f-4ab9-927c-02110a0d74b2-kube-api-access-f5ld5\") pod \"69a1f456-df8f-4ab9-927c-02110a0d74b2\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " Dec 05 11:04:26.734715 master-0 kubenswrapper[24928]: I1205 11:04:26.715602 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-dns-svc\") pod \"69a1f456-df8f-4ab9-927c-02110a0d74b2\" (UID: \"69a1f456-df8f-4ab9-927c-02110a0d74b2\") " Dec 05 11:04:26.734715 master-0 kubenswrapper[24928]: I1205 11:04:26.719764 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69a1f456-df8f-4ab9-927c-02110a0d74b2-kube-api-access-f5ld5" (OuterVolumeSpecName: "kube-api-access-f5ld5") pod "69a1f456-df8f-4ab9-927c-02110a0d74b2" (UID: "69a1f456-df8f-4ab9-927c-02110a0d74b2"). InnerVolumeSpecName "kube-api-access-f5ld5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:26.767040 master-0 kubenswrapper[24928]: I1205 11:04:26.766923 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-config" (OuterVolumeSpecName: "config") pod "69a1f456-df8f-4ab9-927c-02110a0d74b2" (UID: "69a1f456-df8f-4ab9-927c-02110a0d74b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:26.774650 master-0 kubenswrapper[24928]: I1205 11:04:26.774556 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "69a1f456-df8f-4ab9-927c-02110a0d74b2" (UID: "69a1f456-df8f-4ab9-927c-02110a0d74b2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:26.817764 master-0 kubenswrapper[24928]: I1205 11:04:26.817673 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:26.817764 master-0 kubenswrapper[24928]: I1205 11:04:26.817738 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5ld5\" (UniqueName: \"kubernetes.io/projected/69a1f456-df8f-4ab9-927c-02110a0d74b2-kube-api-access-f5ld5\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:26.817764 master-0 kubenswrapper[24928]: I1205 11:04:26.817754 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/69a1f456-df8f-4ab9-927c-02110a0d74b2-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:27.657323 master-0 kubenswrapper[24928]: I1205 11:04:27.657182 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" Dec 05 11:04:27.716316 master-0 kubenswrapper[24928]: I1205 11:04:27.716249 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-658bb5765c-62tp9"] Dec 05 11:04:27.729137 master-0 kubenswrapper[24928]: I1205 11:04:27.729051 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-658bb5765c-62tp9"] Dec 05 11:04:28.224255 master-0 kubenswrapper[24928]: I1205 11:04:28.223870 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69a1f456-df8f-4ab9-927c-02110a0d74b2" path="/var/lib/kubelet/pods/69a1f456-df8f-4ab9-927c-02110a0d74b2/volumes" Dec 05 11:04:28.522783 master-0 kubenswrapper[24928]: I1205 11:04:28.522666 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:04:28.554595 master-0 kubenswrapper[24928]: I1205 11:04:28.554507 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-config\") pod \"02058033-98c5-41b9-8b62-a366ea30a271\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " Dec 05 11:04:28.554877 master-0 kubenswrapper[24928]: I1205 11:04:28.554624 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpxmj\" (UniqueName: \"kubernetes.io/projected/02058033-98c5-41b9-8b62-a366ea30a271-kube-api-access-zpxmj\") pod \"02058033-98c5-41b9-8b62-a366ea30a271\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " Dec 05 11:04:28.554877 master-0 kubenswrapper[24928]: I1205 11:04:28.554676 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-dns-svc\") pod \"02058033-98c5-41b9-8b62-a366ea30a271\" (UID: \"02058033-98c5-41b9-8b62-a366ea30a271\") " Dec 05 11:04:28.596102 master-0 kubenswrapper[24928]: I1205 11:04:28.596047 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02058033-98c5-41b9-8b62-a366ea30a271-kube-api-access-zpxmj" (OuterVolumeSpecName: "kube-api-access-zpxmj") pod "02058033-98c5-41b9-8b62-a366ea30a271" (UID: "02058033-98c5-41b9-8b62-a366ea30a271"). InnerVolumeSpecName "kube-api-access-zpxmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:28.628267 master-0 kubenswrapper[24928]: I1205 11:04:28.628137 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-config" (OuterVolumeSpecName: "config") pod "02058033-98c5-41b9-8b62-a366ea30a271" (UID: "02058033-98c5-41b9-8b62-a366ea30a271"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:28.629627 master-0 kubenswrapper[24928]: I1205 11:04:28.629556 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "02058033-98c5-41b9-8b62-a366ea30a271" (UID: "02058033-98c5-41b9-8b62-a366ea30a271"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:28.660677 master-0 kubenswrapper[24928]: I1205 11:04:28.659716 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:28.660677 master-0 kubenswrapper[24928]: I1205 11:04:28.659794 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpxmj\" (UniqueName: \"kubernetes.io/projected/02058033-98c5-41b9-8b62-a366ea30a271-kube-api-access-zpxmj\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:28.660677 master-0 kubenswrapper[24928]: I1205 11:04:28.659809 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/02058033-98c5-41b9-8b62-a366ea30a271-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:28.674394 master-0 kubenswrapper[24928]: I1205 11:04:28.674328 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" event={"ID":"02058033-98c5-41b9-8b62-a366ea30a271","Type":"ContainerDied","Data":"a7c9b5b3df77acc1dc86ca41be49346bc3fb1e64a82e2e2dc19367f9e7550053"} Dec 05 11:04:28.674630 master-0 kubenswrapper[24928]: I1205 11:04:28.674404 24928 scope.go:117] "RemoveContainer" containerID="bc019b4bff4412a64eba2e8ff6fcfccbb7824ee0c88b30e066d46504c8da6e47" Dec 05 11:04:28.674681 master-0 kubenswrapper[24928]: I1205 11:04:28.674629 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f856ff86c-mztch" Dec 05 11:04:28.711586 master-0 kubenswrapper[24928]: I1205 11:04:28.711527 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f856ff86c-mztch"] Dec 05 11:04:28.793832 master-0 kubenswrapper[24928]: I1205 11:04:28.793775 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f856ff86c-mztch"] Dec 05 11:04:29.452065 master-0 kubenswrapper[24928]: I1205 11:04:29.449985 24928 scope.go:117] "RemoveContainer" containerID="c336e09829a33dc28360bc763604848c019dde17a8f26a6a500676279632363c" Dec 05 11:04:30.227106 master-0 kubenswrapper[24928]: I1205 11:04:30.227049 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02058033-98c5-41b9-8b62-a366ea30a271" path="/var/lib/kubelet/pods/02058033-98c5-41b9-8b62-a366ea30a271/volumes" Dec 05 11:04:30.562190 master-0 kubenswrapper[24928]: I1205 11:04:30.562111 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-658bb5765c-62tp9" podUID="69a1f456-df8f-4ab9-927c-02110a0d74b2" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.128.0.164:5353: i/o timeout" Dec 05 11:04:31.725491 master-0 kubenswrapper[24928]: I1205 11:04:31.725125 24928 generic.go:334] "Generic (PLEG): container finished" podID="5337585b-1757-4276-99fc-672cf04a1a89" containerID="2f405e7bb026d757ed7ec977a2c0bb77e68eae309598e289468fc835898db88c" exitCode=0 Dec 05 11:04:31.725491 master-0 kubenswrapper[24928]: I1205 11:04:31.725250 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" event={"ID":"5337585b-1757-4276-99fc-672cf04a1a89","Type":"ContainerDied","Data":"2f405e7bb026d757ed7ec977a2c0bb77e68eae309598e289468fc835898db88c"} Dec 05 11:04:31.733102 master-0 kubenswrapper[24928]: I1205 11:04:31.733048 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rvm7b" event={"ID":"40924985-cc8e-49f9-9215-8aca88c585ee","Type":"ContainerStarted","Data":"44f1847e217d5b5263b6ba0b3d838ea964241c5779694390298ccfa15ec8ce7a"} Dec 05 11:04:31.738090 master-0 kubenswrapper[24928]: I1205 11:04:31.738029 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a3d90bd2-4fb8-4f48-abe2-5eb0915af549","Type":"ContainerStarted","Data":"3ecbcdd0f1c84cc7b07cfb495e40e834fd30c41a25c96277aacc9a71371112fb"} Dec 05 11:04:31.742194 master-0 kubenswrapper[24928]: I1205 11:04:31.740739 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92","Type":"ContainerStarted","Data":"be690dc852aab0ad22381f0eaec44fa135e8739e2967ec32372d4375d444ed12"} Dec 05 11:04:31.751608 master-0 kubenswrapper[24928]: I1205 11:04:31.751530 24928 generic.go:334] "Generic (PLEG): container finished" podID="7fd3da2c-d0c3-439a-8867-de018998db88" containerID="37cd0810358922b393a436f10e5de4068a383e25a97be24d99afae8dd338af71" exitCode=0 Dec 05 11:04:31.751908 master-0 kubenswrapper[24928]: I1205 11:04:31.751709 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-669479795c-v7gc5" event={"ID":"7fd3da2c-d0c3-439a-8867-de018998db88","Type":"ContainerDied","Data":"37cd0810358922b393a436f10e5de4068a383e25a97be24d99afae8dd338af71"} Dec 05 11:04:31.761244 master-0 kubenswrapper[24928]: I1205 11:04:31.761193 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"51ecb4c8-aa75-4e4f-8529-4492445a9bb4","Type":"ContainerStarted","Data":"330ceede9a3bf36c58fa228a39d071084f463d5be04bc0c085735e4c67caa6e2"} Dec 05 11:04:31.789353 master-0 kubenswrapper[24928]: I1205 11:04:31.783475 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjbjl" event={"ID":"6d62e2eb-0395-4f42-9371-1d6ea1df3a09","Type":"ContainerStarted","Data":"5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d"} Dec 05 11:04:31.795716 master-0 kubenswrapper[24928]: I1205 11:04:31.792495 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"a177c607-c2db-411d-b4ee-e714452fea56","Type":"ContainerStarted","Data":"03219322ff0954390010e939547ed600fbec244999444f3ea30321d7aa5a774c"} Dec 05 11:04:31.795716 master-0 kubenswrapper[24928]: I1205 11:04:31.792706 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 05 11:04:31.929979 master-0 kubenswrapper[24928]: I1205 11:04:31.929633 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=19.076125869 podStartE2EDuration="28.929617398s" podCreationTimestamp="2025-12-05 11:04:03 +0000 UTC" firstStartedPulling="2025-12-05 11:04:19.642143613 +0000 UTC m=+1019.645337464" lastFinishedPulling="2025-12-05 11:04:29.495635142 +0000 UTC m=+1029.498828993" observedRunningTime="2025-12-05 11:04:31.901279931 +0000 UTC m=+1031.904473782" watchObservedRunningTime="2025-12-05 11:04:31.929617398 +0000 UTC m=+1031.932811249" Dec 05 11:04:32.805778 master-0 kubenswrapper[24928]: I1205 11:04:32.805702 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-669479795c-v7gc5" event={"ID":"7fd3da2c-d0c3-439a-8867-de018998db88","Type":"ContainerStarted","Data":"3a22b1ef4250075eaf588a4cfe46a335f8860c202a5a80b69f54d9361af44736"} Dec 05 11:04:32.806303 master-0 kubenswrapper[24928]: I1205 11:04:32.805884 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:32.808393 master-0 kubenswrapper[24928]: I1205 11:04:32.808345 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-bjmv9" event={"ID":"8f43649b-0385-4470-92ec-b031bb81f079","Type":"ContainerStarted","Data":"8f56d3ba9326ed16836e3aca7121b15ea43f54f6700f586bbd9b171ca83969b7"} Dec 05 11:04:32.810803 master-0 kubenswrapper[24928]: I1205 11:04:32.810756 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"51ecb4c8-aa75-4e4f-8529-4492445a9bb4","Type":"ContainerStarted","Data":"372efe9d05fce805600cfb8109707161bf430c6c1d1dd94cc6d7ed4adf8f327b"} Dec 05 11:04:32.813625 master-0 kubenswrapper[24928]: I1205 11:04:32.813571 24928 generic.go:334] "Generic (PLEG): container finished" podID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" containerID="5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d" exitCode=0 Dec 05 11:04:32.813727 master-0 kubenswrapper[24928]: I1205 11:04:32.813643 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjbjl" event={"ID":"6d62e2eb-0395-4f42-9371-1d6ea1df3a09","Type":"ContainerDied","Data":"5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d"} Dec 05 11:04:32.816020 master-0 kubenswrapper[24928]: I1205 11:04:32.815971 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rz85b" event={"ID":"24ba60fc-8a32-4d40-973f-ccabca97d566","Type":"ContainerStarted","Data":"11f76e4cb2c409028bb89d78787f229be0d54b4174302f5e93872b2fd3ceb2be"} Dec 05 11:04:32.816699 master-0 kubenswrapper[24928]: I1205 11:04:32.816671 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-rz85b" Dec 05 11:04:32.819414 master-0 kubenswrapper[24928]: I1205 11:04:32.819372 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"75506b30-c0c5-4cfd-b151-3dbe75a05a77","Type":"ContainerStarted","Data":"054a08031a5842d59e65f3678803984919de55ee34c60bb0b77e2e094235f9e6"} Dec 05 11:04:32.819414 master-0 kubenswrapper[24928]: I1205 11:04:32.819409 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"75506b30-c0c5-4cfd-b151-3dbe75a05a77","Type":"ContainerStarted","Data":"8cd7d43c174ec47062e79fe6f4ca9c0ebea380e507d1a2df20ff1d27e31e6dbf"} Dec 05 11:04:32.826030 master-0 kubenswrapper[24928]: I1205 11:04:32.825958 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" event={"ID":"5337585b-1757-4276-99fc-672cf04a1a89","Type":"ContainerStarted","Data":"06f90d83e4631f295c3fb423c8167bcab0f2e9c635020dbf8a539ff77f63ff64"} Dec 05 11:04:32.826925 master-0 kubenswrapper[24928]: I1205 11:04:32.826869 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:32.841803 master-0 kubenswrapper[24928]: I1205 11:04:32.841732 24928 generic.go:334] "Generic (PLEG): container finished" podID="40924985-cc8e-49f9-9215-8aca88c585ee" containerID="44f1847e217d5b5263b6ba0b3d838ea964241c5779694390298ccfa15ec8ce7a" exitCode=0 Dec 05 11:04:32.842607 master-0 kubenswrapper[24928]: I1205 11:04:32.842546 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rvm7b" event={"ID":"40924985-cc8e-49f9-9215-8aca88c585ee","Type":"ContainerDied","Data":"44f1847e217d5b5263b6ba0b3d838ea964241c5779694390298ccfa15ec8ce7a"} Dec 05 11:04:32.876812 master-0 kubenswrapper[24928]: I1205 11:04:32.876719 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-669479795c-v7gc5" podStartSLOduration=11.876701069 podStartE2EDuration="11.876701069s" podCreationTimestamp="2025-12-05 11:04:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:04:32.844081985 +0000 UTC m=+1032.847275846" watchObservedRunningTime="2025-12-05 11:04:32.876701069 +0000 UTC m=+1032.879894910" Dec 05 11:04:32.883005 master-0 kubenswrapper[24928]: I1205 11:04:32.882907 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rz85b" podStartSLOduration=13.878031889 podStartE2EDuration="23.882882563s" podCreationTimestamp="2025-12-05 11:04:09 +0000 UTC" firstStartedPulling="2025-12-05 11:04:19.642045321 +0000 UTC m=+1019.645239172" lastFinishedPulling="2025-12-05 11:04:29.646895995 +0000 UTC m=+1029.650089846" observedRunningTime="2025-12-05 11:04:32.87393387 +0000 UTC m=+1032.877127721" watchObservedRunningTime="2025-12-05 11:04:32.882882563 +0000 UTC m=+1032.886076414" Dec 05 11:04:32.900539 master-0 kubenswrapper[24928]: I1205 11:04:32.900413 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" podStartSLOduration=11.90039196 podStartE2EDuration="11.90039196s" podCreationTimestamp="2025-12-05 11:04:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:04:32.894507833 +0000 UTC m=+1032.897701684" watchObservedRunningTime="2025-12-05 11:04:32.90039196 +0000 UTC m=+1032.903585821" Dec 05 11:04:32.954039 master-0 kubenswrapper[24928]: I1205 11:04:32.952582 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=14.834631111 podStartE2EDuration="22.95252366s" podCreationTimestamp="2025-12-05 11:04:10 +0000 UTC" firstStartedPulling="2025-12-05 11:04:21.527385245 +0000 UTC m=+1021.530579096" lastFinishedPulling="2025-12-05 11:04:29.645277754 +0000 UTC m=+1029.648471645" observedRunningTime="2025-12-05 11:04:32.948849218 +0000 UTC m=+1032.952043069" watchObservedRunningTime="2025-12-05 11:04:32.95252366 +0000 UTC m=+1032.955717511" Dec 05 11:04:32.987080 master-0 kubenswrapper[24928]: I1205 11:04:32.986439 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=12.988429127 podStartE2EDuration="20.986401785s" podCreationTimestamp="2025-12-05 11:04:12 +0000 UTC" firstStartedPulling="2025-12-05 11:04:21.521396206 +0000 UTC m=+1021.524590057" lastFinishedPulling="2025-12-05 11:04:29.519368864 +0000 UTC m=+1029.522562715" observedRunningTime="2025-12-05 11:04:32.983080563 +0000 UTC m=+1032.986274414" watchObservedRunningTime="2025-12-05 11:04:32.986401785 +0000 UTC m=+1032.989595646" Dec 05 11:04:33.010272 master-0 kubenswrapper[24928]: I1205 11:04:33.010161 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-bjmv9" podStartSLOduration=3.214965986 podStartE2EDuration="12.010139807s" podCreationTimestamp="2025-12-05 11:04:21 +0000 UTC" firstStartedPulling="2025-12-05 11:04:22.504607468 +0000 UTC m=+1022.507801329" lastFinishedPulling="2025-12-05 11:04:31.299781299 +0000 UTC m=+1031.302975150" observedRunningTime="2025-12-05 11:04:33.007513631 +0000 UTC m=+1033.010707482" watchObservedRunningTime="2025-12-05 11:04:33.010139807 +0000 UTC m=+1033.013333658" Dec 05 11:04:33.884269 master-0 kubenswrapper[24928]: I1205 11:04:33.884202 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjbjl" event={"ID":"6d62e2eb-0395-4f42-9371-1d6ea1df3a09","Type":"ContainerStarted","Data":"95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21"} Dec 05 11:04:33.888159 master-0 kubenswrapper[24928]: I1205 11:04:33.888080 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rvm7b" event={"ID":"40924985-cc8e-49f9-9215-8aca88c585ee","Type":"ContainerStarted","Data":"df9bd807e797886369b866eeaaf4d7d30ef10a2e879c7ade32c90459cf7b978a"} Dec 05 11:04:33.888159 master-0 kubenswrapper[24928]: I1205 11:04:33.888130 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-rvm7b" event={"ID":"40924985-cc8e-49f9-9215-8aca88c585ee","Type":"ContainerStarted","Data":"818b23ca58fd610be923fd21d5c701604323a37e0b91e7e18e2962262330fa39"} Dec 05 11:04:33.921491 master-0 kubenswrapper[24928]: I1205 11:04:33.918010 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pjbjl" podStartSLOduration=6.393256482 podStartE2EDuration="16.91798412s" podCreationTimestamp="2025-12-05 11:04:17 +0000 UTC" firstStartedPulling="2025-12-05 11:04:22.574541873 +0000 UTC m=+1022.577735724" lastFinishedPulling="2025-12-05 11:04:33.099269511 +0000 UTC m=+1033.102463362" observedRunningTime="2025-12-05 11:04:33.903739444 +0000 UTC m=+1033.906933305" watchObservedRunningTime="2025-12-05 11:04:33.91798412 +0000 UTC m=+1033.921177991" Dec 05 11:04:33.940698 master-0 kubenswrapper[24928]: I1205 11:04:33.940592 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-rvm7b" podStartSLOduration=15.312026855 podStartE2EDuration="24.940544633s" podCreationTimestamp="2025-12-05 11:04:09 +0000 UTC" firstStartedPulling="2025-12-05 11:04:19.874082378 +0000 UTC m=+1019.877276229" lastFinishedPulling="2025-12-05 11:04:29.502600156 +0000 UTC m=+1029.505794007" observedRunningTime="2025-12-05 11:04:33.927705112 +0000 UTC m=+1033.930898973" watchObservedRunningTime="2025-12-05 11:04:33.940544633 +0000 UTC m=+1033.943738494" Dec 05 11:04:34.120455 master-0 kubenswrapper[24928]: I1205 11:04:34.120331 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:34.120455 master-0 kubenswrapper[24928]: I1205 11:04:34.120401 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:34.177959 master-0 kubenswrapper[24928]: I1205 11:04:34.177885 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:34.733259 master-0 kubenswrapper[24928]: I1205 11:04:34.733198 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:34.733259 master-0 kubenswrapper[24928]: I1205 11:04:34.733252 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:04:35.469276 master-0 kubenswrapper[24928]: I1205 11:04:35.469181 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:35.512272 master-0 kubenswrapper[24928]: I1205 11:04:35.512187 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:35.913634 master-0 kubenswrapper[24928]: I1205 11:04:35.913537 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:36.206940 master-0 kubenswrapper[24928]: E1205 11:04:36.206801 24928 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd7fe9b2_e51f_4f2e_b1f1_ad8f92d7ba92.slice/crio-conmon-be690dc852aab0ad22381f0eaec44fa135e8739e2967ec32372d4375d444ed12.scope\": RecentStats: unable to find data in memory cache]" Dec 05 11:04:36.923573 master-0 kubenswrapper[24928]: I1205 11:04:36.923512 24928 generic.go:334] "Generic (PLEG): container finished" podID="cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92" containerID="be690dc852aab0ad22381f0eaec44fa135e8739e2967ec32372d4375d444ed12" exitCode=0 Dec 05 11:04:36.924218 master-0 kubenswrapper[24928]: I1205 11:04:36.923601 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92","Type":"ContainerDied","Data":"be690dc852aab0ad22381f0eaec44fa135e8739e2967ec32372d4375d444ed12"} Dec 05 11:04:36.926437 master-0 kubenswrapper[24928]: I1205 11:04:36.926375 24928 generic.go:334] "Generic (PLEG): container finished" podID="a3d90bd2-4fb8-4f48-abe2-5eb0915af549" containerID="3ecbcdd0f1c84cc7b07cfb495e40e834fd30c41a25c96277aacc9a71371112fb" exitCode=0 Dec 05 11:04:36.926521 master-0 kubenswrapper[24928]: I1205 11:04:36.926398 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a3d90bd2-4fb8-4f48-abe2-5eb0915af549","Type":"ContainerDied","Data":"3ecbcdd0f1c84cc7b07cfb495e40e834fd30c41a25c96277aacc9a71371112fb"} Dec 05 11:04:36.988256 master-0 kubenswrapper[24928]: I1205 11:04:36.988191 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 05 11:04:37.181561 master-0 kubenswrapper[24928]: I1205 11:04:37.181515 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:37.600915 master-0 kubenswrapper[24928]: I1205 11:04:37.600850 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:37.686710 master-0 kubenswrapper[24928]: I1205 11:04:37.686657 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-669479795c-v7gc5"] Dec 05 11:04:37.767063 master-0 kubenswrapper[24928]: I1205 11:04:37.766972 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:37.767063 master-0 kubenswrapper[24928]: I1205 11:04:37.767058 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:37.819497 master-0 kubenswrapper[24928]: I1205 11:04:37.819415 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:37.945435 master-0 kubenswrapper[24928]: I1205 11:04:37.945216 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a3d90bd2-4fb8-4f48-abe2-5eb0915af549","Type":"ContainerStarted","Data":"abd96dd0a3ee2047c03d68c1c7bedacd133ae79c99d059a01f1ee007a40ba0f8"} Dec 05 11:04:37.950399 master-0 kubenswrapper[24928]: I1205 11:04:37.950350 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"cd7fe9b2-e51f-4f2e-b1f1-ad8f92d7ba92","Type":"ContainerStarted","Data":"b1fb342045117808632ca6982097867697e911dc338231fdce762721d7276108"} Dec 05 11:04:37.951522 master-0 kubenswrapper[24928]: I1205 11:04:37.951482 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-669479795c-v7gc5" podUID="7fd3da2c-d0c3-439a-8867-de018998db88" containerName="dnsmasq-dns" containerID="cri-o://3a22b1ef4250075eaf588a4cfe46a335f8860c202a5a80b69f54d9361af44736" gracePeriod=10 Dec 05 11:04:38.096484 master-0 kubenswrapper[24928]: I1205 11:04:38.095480 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 05 11:04:38.096484 master-0 kubenswrapper[24928]: I1205 11:04:38.095604 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 05 11:04:38.147486 master-0 kubenswrapper[24928]: I1205 11:04:38.140169 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=26.730098376 podStartE2EDuration="37.140123445s" podCreationTimestamp="2025-12-05 11:04:01 +0000 UTC" firstStartedPulling="2025-12-05 11:04:19.235894281 +0000 UTC m=+1019.239088132" lastFinishedPulling="2025-12-05 11:04:29.64591935 +0000 UTC m=+1029.649113201" observedRunningTime="2025-12-05 11:04:38.133657923 +0000 UTC m=+1038.136851774" watchObservedRunningTime="2025-12-05 11:04:38.140123445 +0000 UTC m=+1038.143317316" Dec 05 11:04:38.159741 master-0 kubenswrapper[24928]: I1205 11:04:38.159619 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=25.557048941 podStartE2EDuration="36.159600901s" podCreationTimestamp="2025-12-05 11:04:02 +0000 UTC" firstStartedPulling="2025-12-05 11:04:19.064384484 +0000 UTC m=+1019.067578335" lastFinishedPulling="2025-12-05 11:04:29.666936444 +0000 UTC m=+1029.670130295" observedRunningTime="2025-12-05 11:04:38.157197931 +0000 UTC m=+1038.160391792" watchObservedRunningTime="2025-12-05 11:04:38.159600901 +0000 UTC m=+1038.162794752" Dec 05 11:04:38.972965 master-0 kubenswrapper[24928]: I1205 11:04:38.972876 24928 generic.go:334] "Generic (PLEG): container finished" podID="7fd3da2c-d0c3-439a-8867-de018998db88" containerID="3a22b1ef4250075eaf588a4cfe46a335f8860c202a5a80b69f54d9361af44736" exitCode=0 Dec 05 11:04:38.973541 master-0 kubenswrapper[24928]: I1205 11:04:38.972933 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-669479795c-v7gc5" event={"ID":"7fd3da2c-d0c3-439a-8867-de018998db88","Type":"ContainerDied","Data":"3a22b1ef4250075eaf588a4cfe46a335f8860c202a5a80b69f54d9361af44736"} Dec 05 11:04:38.973541 master-0 kubenswrapper[24928]: I1205 11:04:38.973032 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-669479795c-v7gc5" event={"ID":"7fd3da2c-d0c3-439a-8867-de018998db88","Type":"ContainerDied","Data":"fd7c85ec1f9f9cc7bc9b46a8ed40d3cbb895478edbf33bc2f5c4b8249e08286a"} Dec 05 11:04:38.973541 master-0 kubenswrapper[24928]: I1205 11:04:38.973049 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd7c85ec1f9f9cc7bc9b46a8ed40d3cbb895478edbf33bc2f5c4b8249e08286a" Dec 05 11:04:39.017976 master-0 kubenswrapper[24928]: I1205 11:04:39.017698 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:39.094692 master-0 kubenswrapper[24928]: I1205 11:04:39.094609 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 05 11:04:39.123694 master-0 kubenswrapper[24928]: I1205 11:04:39.123636 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t27hk\" (UniqueName: \"kubernetes.io/projected/7fd3da2c-d0c3-439a-8867-de018998db88-kube-api-access-t27hk\") pod \"7fd3da2c-d0c3-439a-8867-de018998db88\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " Dec 05 11:04:39.123935 master-0 kubenswrapper[24928]: I1205 11:04:39.123720 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-ovsdbserver-nb\") pod \"7fd3da2c-d0c3-439a-8867-de018998db88\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " Dec 05 11:04:39.123935 master-0 kubenswrapper[24928]: I1205 11:04:39.123875 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-config\") pod \"7fd3da2c-d0c3-439a-8867-de018998db88\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " Dec 05 11:04:39.124067 master-0 kubenswrapper[24928]: I1205 11:04:39.124041 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-dns-svc\") pod \"7fd3da2c-d0c3-439a-8867-de018998db88\" (UID: \"7fd3da2c-d0c3-439a-8867-de018998db88\") " Dec 05 11:04:39.140532 master-0 kubenswrapper[24928]: I1205 11:04:39.129791 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fd3da2c-d0c3-439a-8867-de018998db88-kube-api-access-t27hk" (OuterVolumeSpecName: "kube-api-access-t27hk") pod "7fd3da2c-d0c3-439a-8867-de018998db88" (UID: "7fd3da2c-d0c3-439a-8867-de018998db88"). InnerVolumeSpecName "kube-api-access-t27hk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:39.194391 master-0 kubenswrapper[24928]: I1205 11:04:39.194217 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7fd3da2c-d0c3-439a-8867-de018998db88" (UID: "7fd3da2c-d0c3-439a-8867-de018998db88"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:39.199243 master-0 kubenswrapper[24928]: I1205 11:04:39.199069 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 05 11:04:39.227538 master-0 kubenswrapper[24928]: I1205 11:04:39.226935 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t27hk\" (UniqueName: \"kubernetes.io/projected/7fd3da2c-d0c3-439a-8867-de018998db88-kube-api-access-t27hk\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:39.227538 master-0 kubenswrapper[24928]: I1205 11:04:39.227005 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:39.229494 master-0 kubenswrapper[24928]: I1205 11:04:39.227774 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-config" (OuterVolumeSpecName: "config") pod "7fd3da2c-d0c3-439a-8867-de018998db88" (UID: "7fd3da2c-d0c3-439a-8867-de018998db88"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:39.245324 master-0 kubenswrapper[24928]: I1205 11:04:39.245263 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7fd3da2c-d0c3-439a-8867-de018998db88" (UID: "7fd3da2c-d0c3-439a-8867-de018998db88"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:39.332016 master-0 kubenswrapper[24928]: I1205 11:04:39.331233 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:39.332016 master-0 kubenswrapper[24928]: I1205 11:04:39.331281 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7fd3da2c-d0c3-439a-8867-de018998db88-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:39.385758 master-0 kubenswrapper[24928]: I1205 11:04:39.385686 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:39.385758 master-0 kubenswrapper[24928]: I1205 11:04:39.385743 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:39.479738 master-0 kubenswrapper[24928]: I1205 11:04:39.479597 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 05 11:04:39.480242 master-0 kubenswrapper[24928]: E1205 11:04:39.480201 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd3da2c-d0c3-439a-8867-de018998db88" containerName="init" Dec 05 11:04:39.480242 master-0 kubenswrapper[24928]: I1205 11:04:39.480231 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd3da2c-d0c3-439a-8867-de018998db88" containerName="init" Dec 05 11:04:39.480369 master-0 kubenswrapper[24928]: E1205 11:04:39.480276 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02058033-98c5-41b9-8b62-a366ea30a271" containerName="init" Dec 05 11:04:39.480369 master-0 kubenswrapper[24928]: I1205 11:04:39.480287 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="02058033-98c5-41b9-8b62-a366ea30a271" containerName="init" Dec 05 11:04:39.480369 master-0 kubenswrapper[24928]: E1205 11:04:39.480299 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02058033-98c5-41b9-8b62-a366ea30a271" containerName="dnsmasq-dns" Dec 05 11:04:39.480369 master-0 kubenswrapper[24928]: I1205 11:04:39.480308 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="02058033-98c5-41b9-8b62-a366ea30a271" containerName="dnsmasq-dns" Dec 05 11:04:39.480369 master-0 kubenswrapper[24928]: E1205 11:04:39.480333 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd3da2c-d0c3-439a-8867-de018998db88" containerName="dnsmasq-dns" Dec 05 11:04:39.480369 master-0 kubenswrapper[24928]: I1205 11:04:39.480341 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd3da2c-d0c3-439a-8867-de018998db88" containerName="dnsmasq-dns" Dec 05 11:04:39.480369 master-0 kubenswrapper[24928]: E1205 11:04:39.480354 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69a1f456-df8f-4ab9-927c-02110a0d74b2" containerName="dnsmasq-dns" Dec 05 11:04:39.480369 master-0 kubenswrapper[24928]: I1205 11:04:39.480361 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="69a1f456-df8f-4ab9-927c-02110a0d74b2" containerName="dnsmasq-dns" Dec 05 11:04:39.480765 master-0 kubenswrapper[24928]: E1205 11:04:39.480395 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69a1f456-df8f-4ab9-927c-02110a0d74b2" containerName="init" Dec 05 11:04:39.480765 master-0 kubenswrapper[24928]: I1205 11:04:39.480405 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="69a1f456-df8f-4ab9-927c-02110a0d74b2" containerName="init" Dec 05 11:04:39.480765 master-0 kubenswrapper[24928]: I1205 11:04:39.480683 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="69a1f456-df8f-4ab9-927c-02110a0d74b2" containerName="dnsmasq-dns" Dec 05 11:04:39.480765 master-0 kubenswrapper[24928]: I1205 11:04:39.480723 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd3da2c-d0c3-439a-8867-de018998db88" containerName="dnsmasq-dns" Dec 05 11:04:39.480765 master-0 kubenswrapper[24928]: I1205 11:04:39.480746 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="02058033-98c5-41b9-8b62-a366ea30a271" containerName="dnsmasq-dns" Dec 05 11:04:39.483006 master-0 kubenswrapper[24928]: I1205 11:04:39.482103 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 11:04:39.495396 master-0 kubenswrapper[24928]: I1205 11:04:39.495337 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 05 11:04:39.495709 master-0 kubenswrapper[24928]: I1205 11:04:39.495673 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 05 11:04:39.495889 master-0 kubenswrapper[24928]: I1205 11:04:39.495855 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 05 11:04:39.506233 master-0 kubenswrapper[24928]: I1205 11:04:39.506168 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 11:04:39.535466 master-0 kubenswrapper[24928]: I1205 11:04:39.535226 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e2f604-6657-488b-bd45-cd201802ab43-config\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.535466 master-0 kubenswrapper[24928]: I1205 11:04:39.535316 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/67e2f604-6657-488b-bd45-cd201802ab43-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.535742 master-0 kubenswrapper[24928]: I1205 11:04:39.535511 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e2f604-6657-488b-bd45-cd201802ab43-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.535854 master-0 kubenswrapper[24928]: I1205 11:04:39.535802 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67e2f604-6657-488b-bd45-cd201802ab43-scripts\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.536010 master-0 kubenswrapper[24928]: I1205 11:04:39.535960 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl2kd\" (UniqueName: \"kubernetes.io/projected/67e2f604-6657-488b-bd45-cd201802ab43-kube-api-access-zl2kd\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.536054 master-0 kubenswrapper[24928]: I1205 11:04:39.536036 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/67e2f604-6657-488b-bd45-cd201802ab43-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.536120 master-0 kubenswrapper[24928]: I1205 11:04:39.536097 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/67e2f604-6657-488b-bd45-cd201802ab43-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.648455 master-0 kubenswrapper[24928]: I1205 11:04:39.638937 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e2f604-6657-488b-bd45-cd201802ab43-config\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.648455 master-0 kubenswrapper[24928]: I1205 11:04:39.639066 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/67e2f604-6657-488b-bd45-cd201802ab43-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.648455 master-0 kubenswrapper[24928]: I1205 11:04:39.639117 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e2f604-6657-488b-bd45-cd201802ab43-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.648455 master-0 kubenswrapper[24928]: I1205 11:04:39.639193 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67e2f604-6657-488b-bd45-cd201802ab43-scripts\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.648455 master-0 kubenswrapper[24928]: I1205 11:04:39.639278 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl2kd\" (UniqueName: \"kubernetes.io/projected/67e2f604-6657-488b-bd45-cd201802ab43-kube-api-access-zl2kd\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.648455 master-0 kubenswrapper[24928]: I1205 11:04:39.639315 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/67e2f604-6657-488b-bd45-cd201802ab43-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.648455 master-0 kubenswrapper[24928]: I1205 11:04:39.639350 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/67e2f604-6657-488b-bd45-cd201802ab43-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.648455 master-0 kubenswrapper[24928]: I1205 11:04:39.640194 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/67e2f604-6657-488b-bd45-cd201802ab43-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.648455 master-0 kubenswrapper[24928]: I1205 11:04:39.642325 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/67e2f604-6657-488b-bd45-cd201802ab43-config\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.648455 master-0 kubenswrapper[24928]: I1205 11:04:39.643829 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67e2f604-6657-488b-bd45-cd201802ab43-scripts\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.667447 master-0 kubenswrapper[24928]: I1205 11:04:39.662414 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/67e2f604-6657-488b-bd45-cd201802ab43-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.673536 master-0 kubenswrapper[24928]: I1205 11:04:39.670289 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/67e2f604-6657-488b-bd45-cd201802ab43-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.689060 master-0 kubenswrapper[24928]: I1205 11:04:39.688999 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl2kd\" (UniqueName: \"kubernetes.io/projected/67e2f604-6657-488b-bd45-cd201802ab43-kube-api-access-zl2kd\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.690548 master-0 kubenswrapper[24928]: I1205 11:04:39.690508 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67e2f604-6657-488b-bd45-cd201802ab43-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"67e2f604-6657-488b-bd45-cd201802ab43\") " pod="openstack/ovn-northd-0" Dec 05 11:04:39.810024 master-0 kubenswrapper[24928]: I1205 11:04:39.809969 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 11:04:39.995373 master-0 kubenswrapper[24928]: I1205 11:04:39.995305 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-669479795c-v7gc5" Dec 05 11:04:40.060618 master-0 kubenswrapper[24928]: I1205 11:04:40.053839 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-669479795c-v7gc5"] Dec 05 11:04:40.069016 master-0 kubenswrapper[24928]: I1205 11:04:40.068960 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-669479795c-v7gc5"] Dec 05 11:04:40.216208 master-0 kubenswrapper[24928]: I1205 11:04:40.216136 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fd3da2c-d0c3-439a-8867-de018998db88" path="/var/lib/kubelet/pods/7fd3da2c-d0c3-439a-8867-de018998db88/volumes" Dec 05 11:04:41.556212 master-0 kubenswrapper[24928]: I1205 11:04:41.556162 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 11:04:42.033083 master-0 kubenswrapper[24928]: I1205 11:04:42.033031 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"67e2f604-6657-488b-bd45-cd201802ab43","Type":"ContainerStarted","Data":"213989f9cb05728e1e34c4d11baa882009f030e14a13e2738a2e74eee714c04d"} Dec 05 11:04:43.991262 master-0 kubenswrapper[24928]: I1205 11:04:43.991186 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:44.072700 master-0 kubenswrapper[24928]: I1205 11:04:44.072628 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 05 11:04:46.247763 master-0 kubenswrapper[24928]: I1205 11:04:46.247242 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 05 11:04:46.424923 master-0 kubenswrapper[24928]: I1205 11:04:46.424844 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5789dc4cf-qbmxl"] Dec 05 11:04:46.427534 master-0 kubenswrapper[24928]: I1205 11:04:46.427473 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.436063 master-0 kubenswrapper[24928]: I1205 11:04:46.435988 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5789dc4cf-qbmxl"] Dec 05 11:04:46.437632 master-0 kubenswrapper[24928]: I1205 11:04:46.437589 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 05 11:04:46.536757 master-0 kubenswrapper[24928]: I1205 11:04:46.536414 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6l77\" (UniqueName: \"kubernetes.io/projected/52a09227-8731-48e6-9b00-c38d799961f0-kube-api-access-k6l77\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.536757 master-0 kubenswrapper[24928]: I1205 11:04:46.536668 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.536963 master-0 kubenswrapper[24928]: I1205 11:04:46.536892 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-config\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.537144 master-0 kubenswrapper[24928]: I1205 11:04:46.537120 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.537237 master-0 kubenswrapper[24928]: I1205 11:04:46.537205 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-dns-svc\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.639044 master-0 kubenswrapper[24928]: I1205 11:04:46.638970 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.639186 master-0 kubenswrapper[24928]: I1205 11:04:46.639092 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-config\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.639186 master-0 kubenswrapper[24928]: I1205 11:04:46.639171 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.639281 master-0 kubenswrapper[24928]: I1205 11:04:46.639210 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-dns-svc\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.639281 master-0 kubenswrapper[24928]: I1205 11:04:46.639239 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6l77\" (UniqueName: \"kubernetes.io/projected/52a09227-8731-48e6-9b00-c38d799961f0-kube-api-access-k6l77\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.640478 master-0 kubenswrapper[24928]: I1205 11:04:46.640444 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-sb\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.641405 master-0 kubenswrapper[24928]: I1205 11:04:46.641352 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-config\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.642191 master-0 kubenswrapper[24928]: I1205 11:04:46.642124 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-nb\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.642905 master-0 kubenswrapper[24928]: I1205 11:04:46.642868 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-dns-svc\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.666769 master-0 kubenswrapper[24928]: I1205 11:04:46.666633 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6l77\" (UniqueName: \"kubernetes.io/projected/52a09227-8731-48e6-9b00-c38d799961f0-kube-api-access-k6l77\") pod \"dnsmasq-dns-5789dc4cf-qbmxl\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:46.821093 master-0 kubenswrapper[24928]: I1205 11:04:46.821021 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:47.114581 master-0 kubenswrapper[24928]: I1205 11:04:47.113705 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"67e2f604-6657-488b-bd45-cd201802ab43","Type":"ContainerStarted","Data":"f9ada9702463f8accc7be8b594cbf845e579abeaf5c890390d7ab971d2a45661"} Dec 05 11:04:47.114581 master-0 kubenswrapper[24928]: I1205 11:04:47.113780 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"67e2f604-6657-488b-bd45-cd201802ab43","Type":"ContainerStarted","Data":"a06ef3b58e93d6a99fbb484c5be4c9938369e2f2953a13318a4699da76d3240b"} Dec 05 11:04:47.114581 master-0 kubenswrapper[24928]: I1205 11:04:47.114058 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 05 11:04:47.149128 master-0 kubenswrapper[24928]: I1205 11:04:47.149015 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.561616501 podStartE2EDuration="8.148989765s" podCreationTimestamp="2025-12-05 11:04:39 +0000 UTC" firstStartedPulling="2025-12-05 11:04:41.560174105 +0000 UTC m=+1041.563367956" lastFinishedPulling="2025-12-05 11:04:46.147547319 +0000 UTC m=+1046.150741220" observedRunningTime="2025-12-05 11:04:47.147198821 +0000 UTC m=+1047.150392692" watchObservedRunningTime="2025-12-05 11:04:47.148989765 +0000 UTC m=+1047.152183616" Dec 05 11:04:47.358768 master-0 kubenswrapper[24928]: I1205 11:04:47.358692 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5789dc4cf-qbmxl"] Dec 05 11:04:47.373851 master-0 kubenswrapper[24928]: W1205 11:04:47.373806 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52a09227_8731_48e6_9b00_c38d799961f0.slice/crio-592a04a817e5713d439060d1e1e8b48e0127ea7799ca0b4415272c74d43ecc32 WatchSource:0}: Error finding container 592a04a817e5713d439060d1e1e8b48e0127ea7799ca0b4415272c74d43ecc32: Status 404 returned error can't find the container with id 592a04a817e5713d439060d1e1e8b48e0127ea7799ca0b4415272c74d43ecc32 Dec 05 11:04:47.821563 master-0 kubenswrapper[24928]: I1205 11:04:47.821260 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:47.885706 master-0 kubenswrapper[24928]: I1205 11:04:47.885511 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pjbjl"] Dec 05 11:04:48.126482 master-0 kubenswrapper[24928]: I1205 11:04:48.126389 24928 generic.go:334] "Generic (PLEG): container finished" podID="52a09227-8731-48e6-9b00-c38d799961f0" containerID="f8a129e1c3ea5111b73e5f09a16e2e59cc6ca1917594df247773587916a7a399" exitCode=0 Dec 05 11:04:48.127225 master-0 kubenswrapper[24928]: I1205 11:04:48.127171 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" event={"ID":"52a09227-8731-48e6-9b00-c38d799961f0","Type":"ContainerDied","Data":"f8a129e1c3ea5111b73e5f09a16e2e59cc6ca1917594df247773587916a7a399"} Dec 05 11:04:48.127285 master-0 kubenswrapper[24928]: I1205 11:04:48.127232 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" event={"ID":"52a09227-8731-48e6-9b00-c38d799961f0","Type":"ContainerStarted","Data":"592a04a817e5713d439060d1e1e8b48e0127ea7799ca0b4415272c74d43ecc32"} Dec 05 11:04:48.127561 master-0 kubenswrapper[24928]: I1205 11:04:48.127405 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pjbjl" podUID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" containerName="registry-server" containerID="cri-o://95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21" gracePeriod=2 Dec 05 11:04:48.405764 master-0 kubenswrapper[24928]: I1205 11:04:48.405698 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 05 11:04:48.443152 master-0 kubenswrapper[24928]: I1205 11:04:48.442061 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 05 11:04:48.443152 master-0 kubenswrapper[24928]: I1205 11:04:48.442268 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 11:04:48.445785 master-0 kubenswrapper[24928]: I1205 11:04:48.445731 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 05 11:04:48.447725 master-0 kubenswrapper[24928]: I1205 11:04:48.447680 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 05 11:04:48.450819 master-0 kubenswrapper[24928]: I1205 11:04:48.450713 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 05 11:04:48.523295 master-0 kubenswrapper[24928]: I1205 11:04:48.523212 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5fc6011b-6e96-4b50-9ba5-f9c705f1a2d5\" (UniqueName: \"kubernetes.io/csi/topolvm.io^eb41ae07-513e-40ac-9e5b-a778d99e92be\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.523295 master-0 kubenswrapper[24928]: I1205 11:04:48.523277 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twrdf\" (UniqueName: \"kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-kube-api-access-twrdf\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.523605 master-0 kubenswrapper[24928]: I1205 11:04:48.523542 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-lock\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.523816 master-0 kubenswrapper[24928]: I1205 11:04:48.523777 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.523897 master-0 kubenswrapper[24928]: I1205 11:04:48.523834 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-cache\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.625880 master-0 kubenswrapper[24928]: I1205 11:04:48.625802 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5fc6011b-6e96-4b50-9ba5-f9c705f1a2d5\" (UniqueName: \"kubernetes.io/csi/topolvm.io^eb41ae07-513e-40ac-9e5b-a778d99e92be\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.626104 master-0 kubenswrapper[24928]: I1205 11:04:48.625895 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twrdf\" (UniqueName: \"kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-kube-api-access-twrdf\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.626104 master-0 kubenswrapper[24928]: I1205 11:04:48.626056 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-lock\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.626167 master-0 kubenswrapper[24928]: I1205 11:04:48.626111 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.626167 master-0 kubenswrapper[24928]: I1205 11:04:48.626139 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-cache\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.627929 master-0 kubenswrapper[24928]: E1205 11:04:48.627868 24928 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 11:04:48.627929 master-0 kubenswrapper[24928]: E1205 11:04:48.627928 24928 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 11:04:48.628047 master-0 kubenswrapper[24928]: I1205 11:04:48.627987 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-cache\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.628047 master-0 kubenswrapper[24928]: E1205 11:04:48.627989 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift podName:cc6a7a43-f73c-4e31-a61e-de8f4a5957b6 nodeName:}" failed. No retries permitted until 2025-12-05 11:04:49.127968302 +0000 UTC m=+1049.131162153 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift") pod "swift-storage-0" (UID: "cc6a7a43-f73c-4e31-a61e-de8f4a5957b6") : configmap "swift-ring-files" not found Dec 05 11:04:48.628047 master-0 kubenswrapper[24928]: I1205 11:04:48.627880 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-lock\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.629028 master-0 kubenswrapper[24928]: I1205 11:04:48.628993 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:04:48.629083 master-0 kubenswrapper[24928]: I1205 11:04:48.629032 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5fc6011b-6e96-4b50-9ba5-f9c705f1a2d5\" (UniqueName: \"kubernetes.io/csi/topolvm.io^eb41ae07-513e-40ac-9e5b-a778d99e92be\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/df02b82448891026b22666bcad62166584a2fd30988227b3ae65d71bee40128a/globalmount\"" pod="openstack/swift-storage-0" Dec 05 11:04:48.646038 master-0 kubenswrapper[24928]: I1205 11:04:48.645987 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twrdf\" (UniqueName: \"kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-kube-api-access-twrdf\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:48.751475 master-0 kubenswrapper[24928]: I1205 11:04:48.751338 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:48.829729 master-0 kubenswrapper[24928]: I1205 11:04:48.829644 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-utilities\") pod \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " Dec 05 11:04:48.829979 master-0 kubenswrapper[24928]: I1205 11:04:48.829799 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-catalog-content\") pod \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " Dec 05 11:04:48.829979 master-0 kubenswrapper[24928]: I1205 11:04:48.829846 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fzmv\" (UniqueName: \"kubernetes.io/projected/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-kube-api-access-4fzmv\") pod \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\" (UID: \"6d62e2eb-0395-4f42-9371-1d6ea1df3a09\") " Dec 05 11:04:48.830946 master-0 kubenswrapper[24928]: I1205 11:04:48.830887 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-utilities" (OuterVolumeSpecName: "utilities") pod "6d62e2eb-0395-4f42-9371-1d6ea1df3a09" (UID: "6d62e2eb-0395-4f42-9371-1d6ea1df3a09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:04:48.846667 master-0 kubenswrapper[24928]: I1205 11:04:48.846602 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-kube-api-access-4fzmv" (OuterVolumeSpecName: "kube-api-access-4fzmv") pod "6d62e2eb-0395-4f42-9371-1d6ea1df3a09" (UID: "6d62e2eb-0395-4f42-9371-1d6ea1df3a09"). InnerVolumeSpecName "kube-api-access-4fzmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:48.885852 master-0 kubenswrapper[24928]: I1205 11:04:48.885732 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d62e2eb-0395-4f42-9371-1d6ea1df3a09" (UID: "6d62e2eb-0395-4f42-9371-1d6ea1df3a09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:04:48.932873 master-0 kubenswrapper[24928]: I1205 11:04:48.932785 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:48.932873 master-0 kubenswrapper[24928]: I1205 11:04:48.932850 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fzmv\" (UniqueName: \"kubernetes.io/projected/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-kube-api-access-4fzmv\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:48.932873 master-0 kubenswrapper[24928]: I1205 11:04:48.932868 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d62e2eb-0395-4f42-9371-1d6ea1df3a09-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:49.137756 master-0 kubenswrapper[24928]: I1205 11:04:49.137614 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:49.138009 master-0 kubenswrapper[24928]: E1205 11:04:49.137860 24928 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 11:04:49.138009 master-0 kubenswrapper[24928]: E1205 11:04:49.137925 24928 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 11:04:49.138115 master-0 kubenswrapper[24928]: E1205 11:04:49.138012 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift podName:cc6a7a43-f73c-4e31-a61e-de8f4a5957b6 nodeName:}" failed. No retries permitted until 2025-12-05 11:04:50.137961233 +0000 UTC m=+1050.141155094 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift") pod "swift-storage-0" (UID: "cc6a7a43-f73c-4e31-a61e-de8f4a5957b6") : configmap "swift-ring-files" not found Dec 05 11:04:49.140356 master-0 kubenswrapper[24928]: I1205 11:04:49.140261 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" event={"ID":"52a09227-8731-48e6-9b00-c38d799961f0","Type":"ContainerStarted","Data":"83d7a70c489bb4eb9e4bb46585224d92019bcadd4ab5ad3fb37892a3a134feaa"} Dec 05 11:04:49.141630 master-0 kubenswrapper[24928]: I1205 11:04:49.141400 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:49.145204 master-0 kubenswrapper[24928]: I1205 11:04:49.145147 24928 generic.go:334] "Generic (PLEG): container finished" podID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" containerID="95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21" exitCode=0 Dec 05 11:04:49.145328 master-0 kubenswrapper[24928]: I1205 11:04:49.145207 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjbjl" Dec 05 11:04:49.145328 master-0 kubenswrapper[24928]: I1205 11:04:49.145233 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjbjl" event={"ID":"6d62e2eb-0395-4f42-9371-1d6ea1df3a09","Type":"ContainerDied","Data":"95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21"} Dec 05 11:04:49.145328 master-0 kubenswrapper[24928]: I1205 11:04:49.145272 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjbjl" event={"ID":"6d62e2eb-0395-4f42-9371-1d6ea1df3a09","Type":"ContainerDied","Data":"e0ad5630d6f12b9e69ea2daa6c141ae44606e41732559ebf1a5fcea279067b14"} Dec 05 11:04:49.145328 master-0 kubenswrapper[24928]: I1205 11:04:49.145310 24928 scope.go:117] "RemoveContainer" containerID="95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21" Dec 05 11:04:49.178231 master-0 kubenswrapper[24928]: I1205 11:04:49.178175 24928 scope.go:117] "RemoveContainer" containerID="5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d" Dec 05 11:04:49.208526 master-0 kubenswrapper[24928]: I1205 11:04:49.194557 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" podStartSLOduration=3.194539993 podStartE2EDuration="3.194539993s" podCreationTimestamp="2025-12-05 11:04:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:04:49.189955439 +0000 UTC m=+1049.193149290" watchObservedRunningTime="2025-12-05 11:04:49.194539993 +0000 UTC m=+1049.197733844" Dec 05 11:04:49.229736 master-0 kubenswrapper[24928]: I1205 11:04:49.219081 24928 scope.go:117] "RemoveContainer" containerID="209ae81a830f495c81f3d5f4b06cb324a020f5e0955cc58e03abb9233bf4bb8c" Dec 05 11:04:49.229736 master-0 kubenswrapper[24928]: I1205 11:04:49.228744 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pjbjl"] Dec 05 11:04:49.247507 master-0 kubenswrapper[24928]: I1205 11:04:49.245668 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pjbjl"] Dec 05 11:04:49.270948 master-0 kubenswrapper[24928]: I1205 11:04:49.270887 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-b2rmc"] Dec 05 11:04:49.271874 master-0 kubenswrapper[24928]: E1205 11:04:49.271849 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" containerName="registry-server" Dec 05 11:04:49.271988 master-0 kubenswrapper[24928]: I1205 11:04:49.271974 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" containerName="registry-server" Dec 05 11:04:49.272104 master-0 kubenswrapper[24928]: E1205 11:04:49.272088 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" containerName="extract-content" Dec 05 11:04:49.272192 master-0 kubenswrapper[24928]: I1205 11:04:49.272175 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" containerName="extract-content" Dec 05 11:04:49.272276 master-0 kubenswrapper[24928]: E1205 11:04:49.272265 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" containerName="extract-utilities" Dec 05 11:04:49.272339 master-0 kubenswrapper[24928]: I1205 11:04:49.272329 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" containerName="extract-utilities" Dec 05 11:04:49.272695 master-0 kubenswrapper[24928]: I1205 11:04:49.272678 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" containerName="registry-server" Dec 05 11:04:49.273628 master-0 kubenswrapper[24928]: I1205 11:04:49.273610 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.280059 master-0 kubenswrapper[24928]: I1205 11:04:49.276214 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 05 11:04:49.280059 master-0 kubenswrapper[24928]: I1205 11:04:49.276793 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 05 11:04:49.280903 master-0 kubenswrapper[24928]: I1205 11:04:49.280859 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 05 11:04:49.305383 master-0 kubenswrapper[24928]: I1205 11:04:49.305336 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-b2rmc"] Dec 05 11:04:49.333808 master-0 kubenswrapper[24928]: I1205 11:04:49.333639 24928 scope.go:117] "RemoveContainer" containerID="95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21" Dec 05 11:04:49.334189 master-0 kubenswrapper[24928]: E1205 11:04:49.334164 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21\": container with ID starting with 95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21 not found: ID does not exist" containerID="95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21" Dec 05 11:04:49.334306 master-0 kubenswrapper[24928]: I1205 11:04:49.334279 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21"} err="failed to get container status \"95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21\": rpc error: code = NotFound desc = could not find container \"95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21\": container with ID starting with 95be09102d687b136689dad89c79a285efeac6e59c0fadbf730427f4c7258d21 not found: ID does not exist" Dec 05 11:04:49.334378 master-0 kubenswrapper[24928]: I1205 11:04:49.334367 24928 scope.go:117] "RemoveContainer" containerID="5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d" Dec 05 11:04:49.335101 master-0 kubenswrapper[24928]: E1205 11:04:49.335060 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d\": container with ID starting with 5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d not found: ID does not exist" containerID="5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d" Dec 05 11:04:49.335176 master-0 kubenswrapper[24928]: I1205 11:04:49.335116 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d"} err="failed to get container status \"5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d\": rpc error: code = NotFound desc = could not find container \"5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d\": container with ID starting with 5415a5981e6511dccb60de55f4246f2e075457dc7e005e23ad310a8460bc753d not found: ID does not exist" Dec 05 11:04:49.335176 master-0 kubenswrapper[24928]: I1205 11:04:49.335155 24928 scope.go:117] "RemoveContainer" containerID="209ae81a830f495c81f3d5f4b06cb324a020f5e0955cc58e03abb9233bf4bb8c" Dec 05 11:04:49.335557 master-0 kubenswrapper[24928]: E1205 11:04:49.335534 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"209ae81a830f495c81f3d5f4b06cb324a020f5e0955cc58e03abb9233bf4bb8c\": container with ID starting with 209ae81a830f495c81f3d5f4b06cb324a020f5e0955cc58e03abb9233bf4bb8c not found: ID does not exist" containerID="209ae81a830f495c81f3d5f4b06cb324a020f5e0955cc58e03abb9233bf4bb8c" Dec 05 11:04:49.335608 master-0 kubenswrapper[24928]: I1205 11:04:49.335563 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209ae81a830f495c81f3d5f4b06cb324a020f5e0955cc58e03abb9233bf4bb8c"} err="failed to get container status \"209ae81a830f495c81f3d5f4b06cb324a020f5e0955cc58e03abb9233bf4bb8c\": rpc error: code = NotFound desc = could not find container \"209ae81a830f495c81f3d5f4b06cb324a020f5e0955cc58e03abb9233bf4bb8c\": container with ID starting with 209ae81a830f495c81f3d5f4b06cb324a020f5e0955cc58e03abb9233bf4bb8c not found: ID does not exist" Dec 05 11:04:49.342902 master-0 kubenswrapper[24928]: I1205 11:04:49.342844 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-ring-data-devices\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.343042 master-0 kubenswrapper[24928]: I1205 11:04:49.342928 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-scripts\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.343042 master-0 kubenswrapper[24928]: I1205 11:04:49.342979 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-combined-ca-bundle\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.343107 master-0 kubenswrapper[24928]: I1205 11:04:49.343037 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75vzp\" (UniqueName: \"kubernetes.io/projected/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-kube-api-access-75vzp\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.343182 master-0 kubenswrapper[24928]: I1205 11:04:49.343148 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-etc-swift\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.343476 master-0 kubenswrapper[24928]: I1205 11:04:49.343365 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-dispersionconf\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.343968 master-0 kubenswrapper[24928]: I1205 11:04:49.343524 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-swiftconf\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.427200 master-0 kubenswrapper[24928]: I1205 11:04:49.427074 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-28pjp"] Dec 05 11:04:49.429604 master-0 kubenswrapper[24928]: I1205 11:04:49.429547 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-28pjp" Dec 05 11:04:49.445698 master-0 kubenswrapper[24928]: I1205 11:04:49.445635 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-swiftconf\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.446628 master-0 kubenswrapper[24928]: I1205 11:04:49.446589 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-ring-data-devices\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.447399 master-0 kubenswrapper[24928]: I1205 11:04:49.447374 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-scripts\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.447587 master-0 kubenswrapper[24928]: I1205 11:04:49.447570 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-combined-ca-bundle\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.447799 master-0 kubenswrapper[24928]: I1205 11:04:49.447779 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75vzp\" (UniqueName: \"kubernetes.io/projected/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-kube-api-access-75vzp\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.448092 master-0 kubenswrapper[24928]: I1205 11:04:49.448071 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-etc-swift\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.448300 master-0 kubenswrapper[24928]: I1205 11:04:49.448281 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-dispersionconf\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.457030 master-0 kubenswrapper[24928]: I1205 11:04:49.452408 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-ring-data-devices\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.463395 master-0 kubenswrapper[24928]: I1205 11:04:49.463131 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-scripts\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.467687 master-0 kubenswrapper[24928]: I1205 11:04:49.467656 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-swiftconf\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.467812 master-0 kubenswrapper[24928]: I1205 11:04:49.467651 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-dispersionconf\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.468068 master-0 kubenswrapper[24928]: I1205 11:04:49.468038 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-combined-ca-bundle\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.469337 master-0 kubenswrapper[24928]: I1205 11:04:49.469290 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-etc-swift\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.476068 master-0 kubenswrapper[24928]: I1205 11:04:49.476012 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75vzp\" (UniqueName: \"kubernetes.io/projected/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-kube-api-access-75vzp\") pod \"swift-ring-rebalance-b2rmc\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.479791 master-0 kubenswrapper[24928]: I1205 11:04:49.479747 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-2c49-account-create-update-62tjp"] Dec 05 11:04:49.483701 master-0 kubenswrapper[24928]: I1205 11:04:49.483662 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2c49-account-create-update-62tjp" Dec 05 11:04:49.485471 master-0 kubenswrapper[24928]: I1205 11:04:49.485433 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 05 11:04:49.491867 master-0 kubenswrapper[24928]: I1205 11:04:49.490360 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-28pjp"] Dec 05 11:04:49.502924 master-0 kubenswrapper[24928]: I1205 11:04:49.502863 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2c49-account-create-update-62tjp"] Dec 05 11:04:49.550283 master-0 kubenswrapper[24928]: I1205 11:04:49.550211 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-operator-scripts\") pod \"glance-db-create-28pjp\" (UID: \"c3af15ab-5fe0-4536-b61f-77c795f6e7f2\") " pod="openstack/glance-db-create-28pjp" Dec 05 11:04:49.550806 master-0 kubenswrapper[24928]: I1205 11:04:49.550782 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbxf4\" (UniqueName: \"kubernetes.io/projected/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-kube-api-access-gbxf4\") pod \"glance-db-create-28pjp\" (UID: \"c3af15ab-5fe0-4536-b61f-77c795f6e7f2\") " pod="openstack/glance-db-create-28pjp" Dec 05 11:04:49.551018 master-0 kubenswrapper[24928]: I1205 11:04:49.550993 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00a4705-aba4-47f1-8a91-349563989651-operator-scripts\") pod \"glance-2c49-account-create-update-62tjp\" (UID: \"e00a4705-aba4-47f1-8a91-349563989651\") " pod="openstack/glance-2c49-account-create-update-62tjp" Dec 05 11:04:49.551357 master-0 kubenswrapper[24928]: I1205 11:04:49.551295 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccn7v\" (UniqueName: \"kubernetes.io/projected/e00a4705-aba4-47f1-8a91-349563989651-kube-api-access-ccn7v\") pod \"glance-2c49-account-create-update-62tjp\" (UID: \"e00a4705-aba4-47f1-8a91-349563989651\") " pod="openstack/glance-2c49-account-create-update-62tjp" Dec 05 11:04:49.623446 master-0 kubenswrapper[24928]: I1205 11:04:49.621943 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:04:49.655034 master-0 kubenswrapper[24928]: I1205 11:04:49.654968 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccn7v\" (UniqueName: \"kubernetes.io/projected/e00a4705-aba4-47f1-8a91-349563989651-kube-api-access-ccn7v\") pod \"glance-2c49-account-create-update-62tjp\" (UID: \"e00a4705-aba4-47f1-8a91-349563989651\") " pod="openstack/glance-2c49-account-create-update-62tjp" Dec 05 11:04:49.655212 master-0 kubenswrapper[24928]: I1205 11:04:49.655091 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-operator-scripts\") pod \"glance-db-create-28pjp\" (UID: \"c3af15ab-5fe0-4536-b61f-77c795f6e7f2\") " pod="openstack/glance-db-create-28pjp" Dec 05 11:04:49.655212 master-0 kubenswrapper[24928]: I1205 11:04:49.655168 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbxf4\" (UniqueName: \"kubernetes.io/projected/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-kube-api-access-gbxf4\") pod \"glance-db-create-28pjp\" (UID: \"c3af15ab-5fe0-4536-b61f-77c795f6e7f2\") " pod="openstack/glance-db-create-28pjp" Dec 05 11:04:49.655310 master-0 kubenswrapper[24928]: I1205 11:04:49.655246 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00a4705-aba4-47f1-8a91-349563989651-operator-scripts\") pod \"glance-2c49-account-create-update-62tjp\" (UID: \"e00a4705-aba4-47f1-8a91-349563989651\") " pod="openstack/glance-2c49-account-create-update-62tjp" Dec 05 11:04:49.656040 master-0 kubenswrapper[24928]: I1205 11:04:49.656007 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00a4705-aba4-47f1-8a91-349563989651-operator-scripts\") pod \"glance-2c49-account-create-update-62tjp\" (UID: \"e00a4705-aba4-47f1-8a91-349563989651\") " pod="openstack/glance-2c49-account-create-update-62tjp" Dec 05 11:04:49.656386 master-0 kubenswrapper[24928]: I1205 11:04:49.656360 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-operator-scripts\") pod \"glance-db-create-28pjp\" (UID: \"c3af15ab-5fe0-4536-b61f-77c795f6e7f2\") " pod="openstack/glance-db-create-28pjp" Dec 05 11:04:49.704536 master-0 kubenswrapper[24928]: I1205 11:04:49.682187 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbxf4\" (UniqueName: \"kubernetes.io/projected/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-kube-api-access-gbxf4\") pod \"glance-db-create-28pjp\" (UID: \"c3af15ab-5fe0-4536-b61f-77c795f6e7f2\") " pod="openstack/glance-db-create-28pjp" Dec 05 11:04:49.704536 master-0 kubenswrapper[24928]: I1205 11:04:49.690434 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccn7v\" (UniqueName: \"kubernetes.io/projected/e00a4705-aba4-47f1-8a91-349563989651-kube-api-access-ccn7v\") pod \"glance-2c49-account-create-update-62tjp\" (UID: \"e00a4705-aba4-47f1-8a91-349563989651\") " pod="openstack/glance-2c49-account-create-update-62tjp" Dec 05 11:04:49.841416 master-0 kubenswrapper[24928]: I1205 11:04:49.841372 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-28pjp" Dec 05 11:04:49.862246 master-0 kubenswrapper[24928]: I1205 11:04:49.862180 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2c49-account-create-update-62tjp" Dec 05 11:04:50.003360 master-0 kubenswrapper[24928]: I1205 11:04:50.002249 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5fc6011b-6e96-4b50-9ba5-f9c705f1a2d5\" (UniqueName: \"kubernetes.io/csi/topolvm.io^eb41ae07-513e-40ac-9e5b-a778d99e92be\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:50.155585 master-0 kubenswrapper[24928]: I1205 11:04:50.151883 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-b2rmc"] Dec 05 11:04:50.187883 master-0 kubenswrapper[24928]: I1205 11:04:50.187795 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:50.188353 master-0 kubenswrapper[24928]: E1205 11:04:50.188258 24928 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 11:04:50.188353 master-0 kubenswrapper[24928]: E1205 11:04:50.188288 24928 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 11:04:50.188353 master-0 kubenswrapper[24928]: E1205 11:04:50.188342 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift podName:cc6a7a43-f73c-4e31-a61e-de8f4a5957b6 nodeName:}" failed. No retries permitted until 2025-12-05 11:04:52.18832508 +0000 UTC m=+1052.191518941 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift") pod "swift-storage-0" (UID: "cc6a7a43-f73c-4e31-a61e-de8f4a5957b6") : configmap "swift-ring-files" not found Dec 05 11:04:50.221611 master-0 kubenswrapper[24928]: I1205 11:04:50.221477 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d62e2eb-0395-4f42-9371-1d6ea1df3a09" path="/var/lib/kubelet/pods/6d62e2eb-0395-4f42-9371-1d6ea1df3a09/volumes" Dec 05 11:04:50.315026 master-0 kubenswrapper[24928]: I1205 11:04:50.314928 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-28pjp"] Dec 05 11:04:50.336169 master-0 kubenswrapper[24928]: W1205 11:04:50.336075 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3af15ab_5fe0_4536_b61f_77c795f6e7f2.slice/crio-5b81920dbfff80aa72c692e74a68898b162a7515cbb070b3c369fef181552a4f WatchSource:0}: Error finding container 5b81920dbfff80aa72c692e74a68898b162a7515cbb070b3c369fef181552a4f: Status 404 returned error can't find the container with id 5b81920dbfff80aa72c692e74a68898b162a7515cbb070b3c369fef181552a4f Dec 05 11:04:50.469323 master-0 kubenswrapper[24928]: I1205 11:04:50.469244 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2c49-account-create-update-62tjp"] Dec 05 11:04:50.481445 master-0 kubenswrapper[24928]: W1205 11:04:50.481287 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode00a4705_aba4_47f1_8a91_349563989651.slice/crio-554e7874e23d5cdd3e2cd213fe5cd8adfb7cfe1cac879983b965d0309a11def6 WatchSource:0}: Error finding container 554e7874e23d5cdd3e2cd213fe5cd8adfb7cfe1cac879983b965d0309a11def6: Status 404 returned error can't find the container with id 554e7874e23d5cdd3e2cd213fe5cd8adfb7cfe1cac879983b965d0309a11def6 Dec 05 11:04:51.185976 master-0 kubenswrapper[24928]: I1205 11:04:51.185795 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b2rmc" event={"ID":"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28","Type":"ContainerStarted","Data":"93f80fd6da9673e6e1738d7eda3dcecf4824563cf4a111d3d6e341f19afcfcd5"} Dec 05 11:04:51.187584 master-0 kubenswrapper[24928]: I1205 11:04:51.187539 24928 generic.go:334] "Generic (PLEG): container finished" podID="e00a4705-aba4-47f1-8a91-349563989651" containerID="e2f6e1eb2ed9d64ace14666de64091f35b9f6f07353f42395136dc8d33792453" exitCode=0 Dec 05 11:04:51.187584 master-0 kubenswrapper[24928]: I1205 11:04:51.187590 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2c49-account-create-update-62tjp" event={"ID":"e00a4705-aba4-47f1-8a91-349563989651","Type":"ContainerDied","Data":"e2f6e1eb2ed9d64ace14666de64091f35b9f6f07353f42395136dc8d33792453"} Dec 05 11:04:51.187805 master-0 kubenswrapper[24928]: I1205 11:04:51.187634 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2c49-account-create-update-62tjp" event={"ID":"e00a4705-aba4-47f1-8a91-349563989651","Type":"ContainerStarted","Data":"554e7874e23d5cdd3e2cd213fe5cd8adfb7cfe1cac879983b965d0309a11def6"} Dec 05 11:04:51.189991 master-0 kubenswrapper[24928]: I1205 11:04:51.189958 24928 generic.go:334] "Generic (PLEG): container finished" podID="c3af15ab-5fe0-4536-b61f-77c795f6e7f2" containerID="af05b34019d940a91a41a72ec26ef13b704ff3e06c119a443966f570240d62c2" exitCode=0 Dec 05 11:04:51.191128 master-0 kubenswrapper[24928]: I1205 11:04:51.191085 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-28pjp" event={"ID":"c3af15ab-5fe0-4536-b61f-77c795f6e7f2","Type":"ContainerDied","Data":"af05b34019d940a91a41a72ec26ef13b704ff3e06c119a443966f570240d62c2"} Dec 05 11:04:51.191128 master-0 kubenswrapper[24928]: I1205 11:04:51.191118 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-28pjp" event={"ID":"c3af15ab-5fe0-4536-b61f-77c795f6e7f2","Type":"ContainerStarted","Data":"5b81920dbfff80aa72c692e74a68898b162a7515cbb070b3c369fef181552a4f"} Dec 05 11:04:52.232873 master-0 kubenswrapper[24928]: I1205 11:04:52.232769 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:52.233621 master-0 kubenswrapper[24928]: E1205 11:04:52.233121 24928 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 11:04:52.233621 master-0 kubenswrapper[24928]: E1205 11:04:52.233148 24928 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 11:04:52.233621 master-0 kubenswrapper[24928]: E1205 11:04:52.233208 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift podName:cc6a7a43-f73c-4e31-a61e-de8f4a5957b6 nodeName:}" failed. No retries permitted until 2025-12-05 11:04:56.23318834 +0000 UTC m=+1056.236382191 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift") pod "swift-storage-0" (UID: "cc6a7a43-f73c-4e31-a61e-de8f4a5957b6") : configmap "swift-ring-files" not found Dec 05 11:04:53.142938 master-0 kubenswrapper[24928]: I1205 11:04:53.142884 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2c49-account-create-update-62tjp" Dec 05 11:04:53.216970 master-0 kubenswrapper[24928]: I1205 11:04:53.216888 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2c49-account-create-update-62tjp" event={"ID":"e00a4705-aba4-47f1-8a91-349563989651","Type":"ContainerDied","Data":"554e7874e23d5cdd3e2cd213fe5cd8adfb7cfe1cac879983b965d0309a11def6"} Dec 05 11:04:53.216970 master-0 kubenswrapper[24928]: I1205 11:04:53.216973 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="554e7874e23d5cdd3e2cd213fe5cd8adfb7cfe1cac879983b965d0309a11def6" Dec 05 11:04:53.217257 master-0 kubenswrapper[24928]: I1205 11:04:53.216976 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2c49-account-create-update-62tjp" Dec 05 11:04:53.252995 master-0 kubenswrapper[24928]: I1205 11:04:53.252913 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00a4705-aba4-47f1-8a91-349563989651-operator-scripts\") pod \"e00a4705-aba4-47f1-8a91-349563989651\" (UID: \"e00a4705-aba4-47f1-8a91-349563989651\") " Dec 05 11:04:53.253620 master-0 kubenswrapper[24928]: I1205 11:04:53.253078 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccn7v\" (UniqueName: \"kubernetes.io/projected/e00a4705-aba4-47f1-8a91-349563989651-kube-api-access-ccn7v\") pod \"e00a4705-aba4-47f1-8a91-349563989651\" (UID: \"e00a4705-aba4-47f1-8a91-349563989651\") " Dec 05 11:04:53.253620 master-0 kubenswrapper[24928]: I1205 11:04:53.253541 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e00a4705-aba4-47f1-8a91-349563989651-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e00a4705-aba4-47f1-8a91-349563989651" (UID: "e00a4705-aba4-47f1-8a91-349563989651"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:53.253977 master-0 kubenswrapper[24928]: I1205 11:04:53.253947 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e00a4705-aba4-47f1-8a91-349563989651-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:53.256629 master-0 kubenswrapper[24928]: I1205 11:04:53.256587 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e00a4705-aba4-47f1-8a91-349563989651-kube-api-access-ccn7v" (OuterVolumeSpecName: "kube-api-access-ccn7v") pod "e00a4705-aba4-47f1-8a91-349563989651" (UID: "e00a4705-aba4-47f1-8a91-349563989651"). InnerVolumeSpecName "kube-api-access-ccn7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:53.356470 master-0 kubenswrapper[24928]: I1205 11:04:53.356323 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccn7v\" (UniqueName: \"kubernetes.io/projected/e00a4705-aba4-47f1-8a91-349563989651-kube-api-access-ccn7v\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:53.599539 master-0 kubenswrapper[24928]: I1205 11:04:53.599462 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-vb5lw"] Dec 05 11:04:53.604658 master-0 kubenswrapper[24928]: E1205 11:04:53.600768 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e00a4705-aba4-47f1-8a91-349563989651" containerName="mariadb-account-create-update" Dec 05 11:04:53.604658 master-0 kubenswrapper[24928]: I1205 11:04:53.600804 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e00a4705-aba4-47f1-8a91-349563989651" containerName="mariadb-account-create-update" Dec 05 11:04:53.604658 master-0 kubenswrapper[24928]: I1205 11:04:53.601038 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="e00a4705-aba4-47f1-8a91-349563989651" containerName="mariadb-account-create-update" Dec 05 11:04:53.604658 master-0 kubenswrapper[24928]: I1205 11:04:53.602455 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vb5lw" Dec 05 11:04:53.614565 master-0 kubenswrapper[24928]: I1205 11:04:53.614290 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vb5lw"] Dec 05 11:04:53.670894 master-0 kubenswrapper[24928]: I1205 11:04:53.670815 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zcpv\" (UniqueName: \"kubernetes.io/projected/6592b5de-3923-472c-af99-f7ab0e1caf3f-kube-api-access-4zcpv\") pod \"keystone-db-create-vb5lw\" (UID: \"6592b5de-3923-472c-af99-f7ab0e1caf3f\") " pod="openstack/keystone-db-create-vb5lw" Dec 05 11:04:53.670894 master-0 kubenswrapper[24928]: I1205 11:04:53.670866 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6592b5de-3923-472c-af99-f7ab0e1caf3f-operator-scripts\") pod \"keystone-db-create-vb5lw\" (UID: \"6592b5de-3923-472c-af99-f7ab0e1caf3f\") " pod="openstack/keystone-db-create-vb5lw" Dec 05 11:04:53.741453 master-0 kubenswrapper[24928]: I1205 11:04:53.741312 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-0a14-account-create-update-85wx2"] Dec 05 11:04:53.746451 master-0 kubenswrapper[24928]: I1205 11:04:53.744487 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0a14-account-create-update-85wx2" Dec 05 11:04:53.749641 master-0 kubenswrapper[24928]: I1205 11:04:53.747577 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 05 11:04:53.760379 master-0 kubenswrapper[24928]: I1205 11:04:53.758726 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0a14-account-create-update-85wx2"] Dec 05 11:04:53.784453 master-0 kubenswrapper[24928]: I1205 11:04:53.777854 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zcpv\" (UniqueName: \"kubernetes.io/projected/6592b5de-3923-472c-af99-f7ab0e1caf3f-kube-api-access-4zcpv\") pod \"keystone-db-create-vb5lw\" (UID: \"6592b5de-3923-472c-af99-f7ab0e1caf3f\") " pod="openstack/keystone-db-create-vb5lw" Dec 05 11:04:53.784453 master-0 kubenswrapper[24928]: I1205 11:04:53.777949 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6592b5de-3923-472c-af99-f7ab0e1caf3f-operator-scripts\") pod \"keystone-db-create-vb5lw\" (UID: \"6592b5de-3923-472c-af99-f7ab0e1caf3f\") " pod="openstack/keystone-db-create-vb5lw" Dec 05 11:04:53.784453 master-0 kubenswrapper[24928]: I1205 11:04:53.779133 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6592b5de-3923-472c-af99-f7ab0e1caf3f-operator-scripts\") pod \"keystone-db-create-vb5lw\" (UID: \"6592b5de-3923-472c-af99-f7ab0e1caf3f\") " pod="openstack/keystone-db-create-vb5lw" Dec 05 11:04:53.797580 master-0 kubenswrapper[24928]: I1205 11:04:53.797265 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zcpv\" (UniqueName: \"kubernetes.io/projected/6592b5de-3923-472c-af99-f7ab0e1caf3f-kube-api-access-4zcpv\") pod \"keystone-db-create-vb5lw\" (UID: \"6592b5de-3923-472c-af99-f7ab0e1caf3f\") " pod="openstack/keystone-db-create-vb5lw" Dec 05 11:04:53.880850 master-0 kubenswrapper[24928]: I1205 11:04:53.880639 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-operator-scripts\") pod \"keystone-0a14-account-create-update-85wx2\" (UID: \"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1\") " pod="openstack/keystone-0a14-account-create-update-85wx2" Dec 05 11:04:53.881058 master-0 kubenswrapper[24928]: I1205 11:04:53.880972 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whw7w\" (UniqueName: \"kubernetes.io/projected/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-kube-api-access-whw7w\") pod \"keystone-0a14-account-create-update-85wx2\" (UID: \"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1\") " pod="openstack/keystone-0a14-account-create-update-85wx2" Dec 05 11:04:53.938480 master-0 kubenswrapper[24928]: I1205 11:04:53.938399 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vb5lw" Dec 05 11:04:53.954672 master-0 kubenswrapper[24928]: I1205 11:04:53.954574 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-tr2vj"] Dec 05 11:04:53.956234 master-0 kubenswrapper[24928]: I1205 11:04:53.956197 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-tr2vj" Dec 05 11:04:53.969938 master-0 kubenswrapper[24928]: I1205 11:04:53.965070 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-tr2vj"] Dec 05 11:04:53.983325 master-0 kubenswrapper[24928]: I1205 11:04:53.983235 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whw7w\" (UniqueName: \"kubernetes.io/projected/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-kube-api-access-whw7w\") pod \"keystone-0a14-account-create-update-85wx2\" (UID: \"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1\") " pod="openstack/keystone-0a14-account-create-update-85wx2" Dec 05 11:04:53.983673 master-0 kubenswrapper[24928]: I1205 11:04:53.983364 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-operator-scripts\") pod \"keystone-0a14-account-create-update-85wx2\" (UID: \"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1\") " pod="openstack/keystone-0a14-account-create-update-85wx2" Dec 05 11:04:53.984533 master-0 kubenswrapper[24928]: I1205 11:04:53.984483 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-operator-scripts\") pod \"keystone-0a14-account-create-update-85wx2\" (UID: \"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1\") " pod="openstack/keystone-0a14-account-create-update-85wx2" Dec 05 11:04:54.001085 master-0 kubenswrapper[24928]: I1205 11:04:54.001032 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whw7w\" (UniqueName: \"kubernetes.io/projected/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-kube-api-access-whw7w\") pod \"keystone-0a14-account-create-update-85wx2\" (UID: \"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1\") " pod="openstack/keystone-0a14-account-create-update-85wx2" Dec 05 11:04:54.058785 master-0 kubenswrapper[24928]: I1205 11:04:54.058695 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-75b5-account-create-update-bpqzn"] Dec 05 11:04:54.060579 master-0 kubenswrapper[24928]: I1205 11:04:54.060538 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75b5-account-create-update-bpqzn" Dec 05 11:04:54.063255 master-0 kubenswrapper[24928]: I1205 11:04:54.063207 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 05 11:04:54.071520 master-0 kubenswrapper[24928]: I1205 11:04:54.071430 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75b5-account-create-update-bpqzn"] Dec 05 11:04:54.078540 master-0 kubenswrapper[24928]: I1205 11:04:54.077881 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0a14-account-create-update-85wx2" Dec 05 11:04:54.093670 master-0 kubenswrapper[24928]: I1205 11:04:54.093578 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghqzz\" (UniqueName: \"kubernetes.io/projected/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-kube-api-access-ghqzz\") pod \"placement-db-create-tr2vj\" (UID: \"c1ffd92e-85ff-4cab-a0c5-55968a23bc53\") " pod="openstack/placement-db-create-tr2vj" Dec 05 11:04:54.093880 master-0 kubenswrapper[24928]: I1205 11:04:54.093769 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-operator-scripts\") pod \"placement-db-create-tr2vj\" (UID: \"c1ffd92e-85ff-4cab-a0c5-55968a23bc53\") " pod="openstack/placement-db-create-tr2vj" Dec 05 11:04:54.196800 master-0 kubenswrapper[24928]: I1205 11:04:54.196638 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghqzz\" (UniqueName: \"kubernetes.io/projected/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-kube-api-access-ghqzz\") pod \"placement-db-create-tr2vj\" (UID: \"c1ffd92e-85ff-4cab-a0c5-55968a23bc53\") " pod="openstack/placement-db-create-tr2vj" Dec 05 11:04:54.197186 master-0 kubenswrapper[24928]: I1205 11:04:54.197166 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-operator-scripts\") pod \"placement-db-create-tr2vj\" (UID: \"c1ffd92e-85ff-4cab-a0c5-55968a23bc53\") " pod="openstack/placement-db-create-tr2vj" Dec 05 11:04:54.198271 master-0 kubenswrapper[24928]: I1205 11:04:54.198067 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-operator-scripts\") pod \"placement-db-create-tr2vj\" (UID: \"c1ffd92e-85ff-4cab-a0c5-55968a23bc53\") " pod="openstack/placement-db-create-tr2vj" Dec 05 11:04:54.198781 master-0 kubenswrapper[24928]: I1205 11:04:54.198757 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae071a88-af53-4fb3-93fc-76d8fd311de0-operator-scripts\") pod \"placement-75b5-account-create-update-bpqzn\" (UID: \"ae071a88-af53-4fb3-93fc-76d8fd311de0\") " pod="openstack/placement-75b5-account-create-update-bpqzn" Dec 05 11:04:54.198976 master-0 kubenswrapper[24928]: I1205 11:04:54.198952 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49t8d\" (UniqueName: \"kubernetes.io/projected/ae071a88-af53-4fb3-93fc-76d8fd311de0-kube-api-access-49t8d\") pod \"placement-75b5-account-create-update-bpqzn\" (UID: \"ae071a88-af53-4fb3-93fc-76d8fd311de0\") " pod="openstack/placement-75b5-account-create-update-bpqzn" Dec 05 11:04:54.219630 master-0 kubenswrapper[24928]: I1205 11:04:54.219381 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghqzz\" (UniqueName: \"kubernetes.io/projected/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-kube-api-access-ghqzz\") pod \"placement-db-create-tr2vj\" (UID: \"c1ffd92e-85ff-4cab-a0c5-55968a23bc53\") " pod="openstack/placement-db-create-tr2vj" Dec 05 11:04:54.275993 master-0 kubenswrapper[24928]: I1205 11:04:54.275893 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-tr2vj" Dec 05 11:04:54.301691 master-0 kubenswrapper[24928]: I1205 11:04:54.301620 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae071a88-af53-4fb3-93fc-76d8fd311de0-operator-scripts\") pod \"placement-75b5-account-create-update-bpqzn\" (UID: \"ae071a88-af53-4fb3-93fc-76d8fd311de0\") " pod="openstack/placement-75b5-account-create-update-bpqzn" Dec 05 11:04:54.301998 master-0 kubenswrapper[24928]: I1205 11:04:54.301701 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49t8d\" (UniqueName: \"kubernetes.io/projected/ae071a88-af53-4fb3-93fc-76d8fd311de0-kube-api-access-49t8d\") pod \"placement-75b5-account-create-update-bpqzn\" (UID: \"ae071a88-af53-4fb3-93fc-76d8fd311de0\") " pod="openstack/placement-75b5-account-create-update-bpqzn" Dec 05 11:04:54.302516 master-0 kubenswrapper[24928]: I1205 11:04:54.302360 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae071a88-af53-4fb3-93fc-76d8fd311de0-operator-scripts\") pod \"placement-75b5-account-create-update-bpqzn\" (UID: \"ae071a88-af53-4fb3-93fc-76d8fd311de0\") " pod="openstack/placement-75b5-account-create-update-bpqzn" Dec 05 11:04:54.317480 master-0 kubenswrapper[24928]: I1205 11:04:54.317322 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49t8d\" (UniqueName: \"kubernetes.io/projected/ae071a88-af53-4fb3-93fc-76d8fd311de0-kube-api-access-49t8d\") pod \"placement-75b5-account-create-update-bpqzn\" (UID: \"ae071a88-af53-4fb3-93fc-76d8fd311de0\") " pod="openstack/placement-75b5-account-create-update-bpqzn" Dec 05 11:04:54.397768 master-0 kubenswrapper[24928]: I1205 11:04:54.397622 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75b5-account-create-update-bpqzn" Dec 05 11:04:54.541689 master-0 kubenswrapper[24928]: I1205 11:04:54.541623 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-28pjp" Dec 05 11:04:54.610030 master-0 kubenswrapper[24928]: I1205 11:04:54.609975 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-operator-scripts\") pod \"c3af15ab-5fe0-4536-b61f-77c795f6e7f2\" (UID: \"c3af15ab-5fe0-4536-b61f-77c795f6e7f2\") " Dec 05 11:04:54.610191 master-0 kubenswrapper[24928]: I1205 11:04:54.610054 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbxf4\" (UniqueName: \"kubernetes.io/projected/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-kube-api-access-gbxf4\") pod \"c3af15ab-5fe0-4536-b61f-77c795f6e7f2\" (UID: \"c3af15ab-5fe0-4536-b61f-77c795f6e7f2\") " Dec 05 11:04:54.614105 master-0 kubenswrapper[24928]: I1205 11:04:54.613109 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c3af15ab-5fe0-4536-b61f-77c795f6e7f2" (UID: "c3af15ab-5fe0-4536-b61f-77c795f6e7f2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:54.616556 master-0 kubenswrapper[24928]: I1205 11:04:54.616516 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-kube-api-access-gbxf4" (OuterVolumeSpecName: "kube-api-access-gbxf4") pod "c3af15ab-5fe0-4536-b61f-77c795f6e7f2" (UID: "c3af15ab-5fe0-4536-b61f-77c795f6e7f2"). InnerVolumeSpecName "kube-api-access-gbxf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:54.713385 master-0 kubenswrapper[24928]: I1205 11:04:54.713195 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:54.713385 master-0 kubenswrapper[24928]: I1205 11:04:54.713250 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbxf4\" (UniqueName: \"kubernetes.io/projected/c3af15ab-5fe0-4536-b61f-77c795f6e7f2-kube-api-access-gbxf4\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:54.940841 master-0 kubenswrapper[24928]: I1205 11:04:54.940789 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vb5lw"] Dec 05 11:04:54.951928 master-0 kubenswrapper[24928]: I1205 11:04:54.951880 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0a14-account-create-update-85wx2"] Dec 05 11:04:55.245711 master-0 kubenswrapper[24928]: I1205 11:04:55.245633 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b2rmc" event={"ID":"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28","Type":"ContainerStarted","Data":"4b3ebf5f20bc7fd19b8fd2c7e9c45f05dc8fd24c06c885d0dafccdbf308eb008"} Dec 05 11:04:55.247684 master-0 kubenswrapper[24928]: I1205 11:04:55.247645 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-28pjp" event={"ID":"c3af15ab-5fe0-4536-b61f-77c795f6e7f2","Type":"ContainerDied","Data":"5b81920dbfff80aa72c692e74a68898b162a7515cbb070b3c369fef181552a4f"} Dec 05 11:04:55.247684 master-0 kubenswrapper[24928]: I1205 11:04:55.247679 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b81920dbfff80aa72c692e74a68898b162a7515cbb070b3c369fef181552a4f" Dec 05 11:04:55.247829 master-0 kubenswrapper[24928]: I1205 11:04:55.247729 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-28pjp" Dec 05 11:04:55.249819 master-0 kubenswrapper[24928]: I1205 11:04:55.249756 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0a14-account-create-update-85wx2" event={"ID":"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1","Type":"ContainerStarted","Data":"a9786f040bbecbf9426a8956ca3aee10e8d972ef9f9edde79cb1669f8eeee386"} Dec 05 11:04:55.252376 master-0 kubenswrapper[24928]: I1205 11:04:55.252261 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vb5lw" event={"ID":"6592b5de-3923-472c-af99-f7ab0e1caf3f","Type":"ContainerStarted","Data":"7e41a47bfc7ae80b9d9d0a990f3507e7d5934f474959949298c8f7a2296489a2"} Dec 05 11:04:55.755805 master-0 kubenswrapper[24928]: I1205 11:04:55.755702 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-b2rmc" podStartSLOduration=2.452586712 podStartE2EDuration="6.755677425s" podCreationTimestamp="2025-12-05 11:04:49 +0000 UTC" firstStartedPulling="2025-12-05 11:04:50.160472135 +0000 UTC m=+1050.163665986" lastFinishedPulling="2025-12-05 11:04:54.463562848 +0000 UTC m=+1054.466756699" observedRunningTime="2025-12-05 11:04:55.751778687 +0000 UTC m=+1055.754972538" watchObservedRunningTime="2025-12-05 11:04:55.755677425 +0000 UTC m=+1055.758871286" Dec 05 11:04:55.777198 master-0 kubenswrapper[24928]: W1205 11:04:55.777121 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae071a88_af53_4fb3_93fc_76d8fd311de0.slice/crio-5e37635c63537ba8ff70df6536db52e83e053e659d62485ce8041d7cec18525d WatchSource:0}: Error finding container 5e37635c63537ba8ff70df6536db52e83e053e659d62485ce8041d7cec18525d: Status 404 returned error can't find the container with id 5e37635c63537ba8ff70df6536db52e83e053e659d62485ce8041d7cec18525d Dec 05 11:04:55.785408 master-0 kubenswrapper[24928]: I1205 11:04:55.785365 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-tr2vj"] Dec 05 11:04:55.797199 master-0 kubenswrapper[24928]: I1205 11:04:55.797132 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-75b5-account-create-update-bpqzn"] Dec 05 11:04:56.255307 master-0 kubenswrapper[24928]: I1205 11:04:56.255165 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:04:56.255652 master-0 kubenswrapper[24928]: E1205 11:04:56.255609 24928 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 05 11:04:56.255652 master-0 kubenswrapper[24928]: E1205 11:04:56.255643 24928 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 05 11:04:56.255732 master-0 kubenswrapper[24928]: E1205 11:04:56.255688 24928 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift podName:cc6a7a43-f73c-4e31-a61e-de8f4a5957b6 nodeName:}" failed. No retries permitted until 2025-12-05 11:05:04.255671485 +0000 UTC m=+1064.258865336 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift") pod "swift-storage-0" (UID: "cc6a7a43-f73c-4e31-a61e-de8f4a5957b6") : configmap "swift-ring-files" not found Dec 05 11:04:56.268289 master-0 kubenswrapper[24928]: I1205 11:04:56.268166 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75b5-account-create-update-bpqzn" event={"ID":"ae071a88-af53-4fb3-93fc-76d8fd311de0","Type":"ContainerStarted","Data":"5e37635c63537ba8ff70df6536db52e83e053e659d62485ce8041d7cec18525d"} Dec 05 11:04:56.270230 master-0 kubenswrapper[24928]: I1205 11:04:56.270200 24928 generic.go:334] "Generic (PLEG): container finished" podID="6592b5de-3923-472c-af99-f7ab0e1caf3f" containerID="18d2f406d5d95ce2b4693cd7c1117c8b7d5bd0491c94e207cd7060fae9bbd22d" exitCode=0 Dec 05 11:04:56.270309 master-0 kubenswrapper[24928]: I1205 11:04:56.270257 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vb5lw" event={"ID":"6592b5de-3923-472c-af99-f7ab0e1caf3f","Type":"ContainerDied","Data":"18d2f406d5d95ce2b4693cd7c1117c8b7d5bd0491c94e207cd7060fae9bbd22d"} Dec 05 11:04:56.272896 master-0 kubenswrapper[24928]: I1205 11:04:56.272859 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-tr2vj" event={"ID":"c1ffd92e-85ff-4cab-a0c5-55968a23bc53","Type":"ContainerStarted","Data":"4ca25df915105587da2490f94a3c11f9c9d27da7edb8087c492a8fd63e058999"} Dec 05 11:04:56.274710 master-0 kubenswrapper[24928]: I1205 11:04:56.274685 24928 generic.go:334] "Generic (PLEG): container finished" podID="6a3249e2-8dd4-42e6-b78e-8712b33d7cf1" containerID="fc404e48fb0c4ef209a261dfd42695a5705750c1cc8ff2b11f16be3e127c429e" exitCode=0 Dec 05 11:04:56.274902 master-0 kubenswrapper[24928]: I1205 11:04:56.274739 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0a14-account-create-update-85wx2" event={"ID":"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1","Type":"ContainerDied","Data":"fc404e48fb0c4ef209a261dfd42695a5705750c1cc8ff2b11f16be3e127c429e"} Dec 05 11:04:56.276817 master-0 kubenswrapper[24928]: I1205 11:04:56.276773 24928 generic.go:334] "Generic (PLEG): container finished" podID="d39078ce-d99a-4e13-bbec-6b986c74a797" containerID="caa0367285949fe4ea582abe038e4d12d2481440aca0e346aa8b61fe6ea33fa4" exitCode=0 Dec 05 11:04:56.277673 master-0 kubenswrapper[24928]: I1205 11:04:56.277637 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d39078ce-d99a-4e13-bbec-6b986c74a797","Type":"ContainerDied","Data":"caa0367285949fe4ea582abe038e4d12d2481440aca0e346aa8b61fe6ea33fa4"} Dec 05 11:04:56.823630 master-0 kubenswrapper[24928]: I1205 11:04:56.823552 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:04:56.941150 master-0 kubenswrapper[24928]: I1205 11:04:56.940594 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57545c5d5f-j2xrl"] Dec 05 11:04:56.941150 master-0 kubenswrapper[24928]: I1205 11:04:56.940993 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" podUID="5337585b-1757-4276-99fc-672cf04a1a89" containerName="dnsmasq-dns" containerID="cri-o://06f90d83e4631f295c3fb423c8167bcab0f2e9c635020dbf8a539ff77f63ff64" gracePeriod=10 Dec 05 11:04:57.312930 master-0 kubenswrapper[24928]: I1205 11:04:57.312850 24928 generic.go:334] "Generic (PLEG): container finished" podID="ae071a88-af53-4fb3-93fc-76d8fd311de0" containerID="1cebad37928253659fa3e85f9d986ce4e48e869f5a6564967bb363fec503cb1d" exitCode=0 Dec 05 11:04:57.313175 master-0 kubenswrapper[24928]: I1205 11:04:57.312951 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75b5-account-create-update-bpqzn" event={"ID":"ae071a88-af53-4fb3-93fc-76d8fd311de0","Type":"ContainerDied","Data":"1cebad37928253659fa3e85f9d986ce4e48e869f5a6564967bb363fec503cb1d"} Dec 05 11:04:57.318307 master-0 kubenswrapper[24928]: I1205 11:04:57.318241 24928 generic.go:334] "Generic (PLEG): container finished" podID="ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc" containerID="4704001173b50a7a4e8491dc21bc67559657f4c7792b6c90ee601e27644cfe75" exitCode=0 Dec 05 11:04:57.318415 master-0 kubenswrapper[24928]: I1205 11:04:57.318336 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc","Type":"ContainerDied","Data":"4704001173b50a7a4e8491dc21bc67559657f4c7792b6c90ee601e27644cfe75"} Dec 05 11:04:57.321783 master-0 kubenswrapper[24928]: I1205 11:04:57.321714 24928 generic.go:334] "Generic (PLEG): container finished" podID="5337585b-1757-4276-99fc-672cf04a1a89" containerID="06f90d83e4631f295c3fb423c8167bcab0f2e9c635020dbf8a539ff77f63ff64" exitCode=0 Dec 05 11:04:57.321930 master-0 kubenswrapper[24928]: I1205 11:04:57.321804 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" event={"ID":"5337585b-1757-4276-99fc-672cf04a1a89","Type":"ContainerDied","Data":"06f90d83e4631f295c3fb423c8167bcab0f2e9c635020dbf8a539ff77f63ff64"} Dec 05 11:04:57.331680 master-0 kubenswrapper[24928]: I1205 11:04:57.331269 24928 generic.go:334] "Generic (PLEG): container finished" podID="c1ffd92e-85ff-4cab-a0c5-55968a23bc53" containerID="b38c6de7050c6c6b24d2ae10f75e1b7c84de9556230059b512e58349b6042964" exitCode=0 Dec 05 11:04:57.331680 master-0 kubenswrapper[24928]: I1205 11:04:57.331374 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-tr2vj" event={"ID":"c1ffd92e-85ff-4cab-a0c5-55968a23bc53","Type":"ContainerDied","Data":"b38c6de7050c6c6b24d2ae10f75e1b7c84de9556230059b512e58349b6042964"} Dec 05 11:04:57.337976 master-0 kubenswrapper[24928]: I1205 11:04:57.337406 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d39078ce-d99a-4e13-bbec-6b986c74a797","Type":"ContainerStarted","Data":"1ecd7f9f1376a4a89adc2aeed36b37cb397e48273b6cc5a3ed6e402ad7ec23cb"} Dec 05 11:04:57.338695 master-0 kubenswrapper[24928]: I1205 11:04:57.338638 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:04:57.522388 master-0 kubenswrapper[24928]: I1205 11:04:57.522352 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:57.595624 master-0 kubenswrapper[24928]: I1205 11:04:57.595558 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-nb\") pod \"5337585b-1757-4276-99fc-672cf04a1a89\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " Dec 05 11:04:57.595911 master-0 kubenswrapper[24928]: I1205 11:04:57.595661 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-dns-svc\") pod \"5337585b-1757-4276-99fc-672cf04a1a89\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " Dec 05 11:04:57.595911 master-0 kubenswrapper[24928]: I1205 11:04:57.595777 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-config\") pod \"5337585b-1757-4276-99fc-672cf04a1a89\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " Dec 05 11:04:57.595911 master-0 kubenswrapper[24928]: I1205 11:04:57.595858 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2g8d\" (UniqueName: \"kubernetes.io/projected/5337585b-1757-4276-99fc-672cf04a1a89-kube-api-access-t2g8d\") pod \"5337585b-1757-4276-99fc-672cf04a1a89\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " Dec 05 11:04:57.595911 master-0 kubenswrapper[24928]: I1205 11:04:57.595891 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-sb\") pod \"5337585b-1757-4276-99fc-672cf04a1a89\" (UID: \"5337585b-1757-4276-99fc-672cf04a1a89\") " Dec 05 11:04:57.601848 master-0 kubenswrapper[24928]: I1205 11:04:57.601784 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5337585b-1757-4276-99fc-672cf04a1a89-kube-api-access-t2g8d" (OuterVolumeSpecName: "kube-api-access-t2g8d") pod "5337585b-1757-4276-99fc-672cf04a1a89" (UID: "5337585b-1757-4276-99fc-672cf04a1a89"). InnerVolumeSpecName "kube-api-access-t2g8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:57.662730 master-0 kubenswrapper[24928]: I1205 11:04:57.662662 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5337585b-1757-4276-99fc-672cf04a1a89" (UID: "5337585b-1757-4276-99fc-672cf04a1a89"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:57.673662 master-0 kubenswrapper[24928]: I1205 11:04:57.673575 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-config" (OuterVolumeSpecName: "config") pod "5337585b-1757-4276-99fc-672cf04a1a89" (UID: "5337585b-1757-4276-99fc-672cf04a1a89"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:57.705841 master-0 kubenswrapper[24928]: I1205 11:04:57.705442 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:57.705841 master-0 kubenswrapper[24928]: I1205 11:04:57.705686 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5337585b-1757-4276-99fc-672cf04a1a89" (UID: "5337585b-1757-4276-99fc-672cf04a1a89"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:57.705841 master-0 kubenswrapper[24928]: I1205 11:04:57.705769 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:57.705841 master-0 kubenswrapper[24928]: I1205 11:04:57.705795 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2g8d\" (UniqueName: \"kubernetes.io/projected/5337585b-1757-4276-99fc-672cf04a1a89-kube-api-access-t2g8d\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:57.723778 master-0 kubenswrapper[24928]: I1205 11:04:57.718662 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5337585b-1757-4276-99fc-672cf04a1a89" (UID: "5337585b-1757-4276-99fc-672cf04a1a89"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:57.814595 master-0 kubenswrapper[24928]: I1205 11:04:57.808440 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:57.814595 master-0 kubenswrapper[24928]: I1205 11:04:57.808547 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5337585b-1757-4276-99fc-672cf04a1a89-ovsdbserver-sb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:57.889481 master-0 kubenswrapper[24928]: I1205 11:04:57.887653 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0a14-account-create-update-85wx2" Dec 05 11:04:57.911178 master-0 kubenswrapper[24928]: I1205 11:04:57.911125 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vb5lw" Dec 05 11:04:58.012179 master-0 kubenswrapper[24928]: I1205 11:04:58.012013 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zcpv\" (UniqueName: \"kubernetes.io/projected/6592b5de-3923-472c-af99-f7ab0e1caf3f-kube-api-access-4zcpv\") pod \"6592b5de-3923-472c-af99-f7ab0e1caf3f\" (UID: \"6592b5de-3923-472c-af99-f7ab0e1caf3f\") " Dec 05 11:04:58.012383 master-0 kubenswrapper[24928]: I1205 11:04:58.012199 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whw7w\" (UniqueName: \"kubernetes.io/projected/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-kube-api-access-whw7w\") pod \"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1\" (UID: \"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1\") " Dec 05 11:04:58.012383 master-0 kubenswrapper[24928]: I1205 11:04:58.012274 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-operator-scripts\") pod \"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1\" (UID: \"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1\") " Dec 05 11:04:58.012518 master-0 kubenswrapper[24928]: I1205 11:04:58.012383 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6592b5de-3923-472c-af99-f7ab0e1caf3f-operator-scripts\") pod \"6592b5de-3923-472c-af99-f7ab0e1caf3f\" (UID: \"6592b5de-3923-472c-af99-f7ab0e1caf3f\") " Dec 05 11:04:58.014160 master-0 kubenswrapper[24928]: I1205 11:04:58.014110 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6a3249e2-8dd4-42e6-b78e-8712b33d7cf1" (UID: "6a3249e2-8dd4-42e6-b78e-8712b33d7cf1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:58.015101 master-0 kubenswrapper[24928]: I1205 11:04:58.015062 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6592b5de-3923-472c-af99-f7ab0e1caf3f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6592b5de-3923-472c-af99-f7ab0e1caf3f" (UID: "6592b5de-3923-472c-af99-f7ab0e1caf3f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:58.041140 master-0 kubenswrapper[24928]: I1205 11:04:58.040321 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-kube-api-access-whw7w" (OuterVolumeSpecName: "kube-api-access-whw7w") pod "6a3249e2-8dd4-42e6-b78e-8712b33d7cf1" (UID: "6a3249e2-8dd4-42e6-b78e-8712b33d7cf1"). InnerVolumeSpecName "kube-api-access-whw7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:58.041140 master-0 kubenswrapper[24928]: I1205 11:04:58.040854 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6592b5de-3923-472c-af99-f7ab0e1caf3f-kube-api-access-4zcpv" (OuterVolumeSpecName: "kube-api-access-4zcpv") pod "6592b5de-3923-472c-af99-f7ab0e1caf3f" (UID: "6592b5de-3923-472c-af99-f7ab0e1caf3f"). InnerVolumeSpecName "kube-api-access-4zcpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:58.116388 master-0 kubenswrapper[24928]: I1205 11:04:58.116317 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zcpv\" (UniqueName: \"kubernetes.io/projected/6592b5de-3923-472c-af99-f7ab0e1caf3f-kube-api-access-4zcpv\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:58.116388 master-0 kubenswrapper[24928]: I1205 11:04:58.116371 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whw7w\" (UniqueName: \"kubernetes.io/projected/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-kube-api-access-whw7w\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:58.116388 master-0 kubenswrapper[24928]: I1205 11:04:58.116387 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:58.116388 master-0 kubenswrapper[24928]: I1205 11:04:58.116403 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6592b5de-3923-472c-af99-f7ab0e1caf3f-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:58.184550 master-0 kubenswrapper[24928]: I1205 11:04:58.184214 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=54.011102682 podStartE2EDuration="58.184190205s" podCreationTimestamp="2025-12-05 11:04:00 +0000 UTC" firstStartedPulling="2025-12-05 11:04:17.452122162 +0000 UTC m=+1017.455316013" lastFinishedPulling="2025-12-05 11:04:21.625209685 +0000 UTC m=+1021.628403536" observedRunningTime="2025-12-05 11:04:58.179961639 +0000 UTC m=+1058.183155490" watchObservedRunningTime="2025-12-05 11:04:58.184190205 +0000 UTC m=+1058.187384056" Dec 05 11:04:58.350597 master-0 kubenswrapper[24928]: I1205 11:04:58.350524 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0a14-account-create-update-85wx2" Dec 05 11:04:58.350929 master-0 kubenswrapper[24928]: I1205 11:04:58.350830 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0a14-account-create-update-85wx2" event={"ID":"6a3249e2-8dd4-42e6-b78e-8712b33d7cf1","Type":"ContainerDied","Data":"a9786f040bbecbf9426a8956ca3aee10e8d972ef9f9edde79cb1669f8eeee386"} Dec 05 11:04:58.351147 master-0 kubenswrapper[24928]: I1205 11:04:58.351117 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9786f040bbecbf9426a8956ca3aee10e8d972ef9f9edde79cb1669f8eeee386" Dec 05 11:04:58.354498 master-0 kubenswrapper[24928]: I1205 11:04:58.352913 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba86f0e1-647d-46a5-b5e8-b05dbe25c5dc","Type":"ContainerStarted","Data":"8724575e6a6ed56b889aeb3c9e749037213a4d344428de3b769075a25071a2bf"} Dec 05 11:04:58.354498 master-0 kubenswrapper[24928]: I1205 11:04:58.353188 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 05 11:04:58.355849 master-0 kubenswrapper[24928]: I1205 11:04:58.355794 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" event={"ID":"5337585b-1757-4276-99fc-672cf04a1a89","Type":"ContainerDied","Data":"4658264ef1211933bb6d627e7e5009cc5963b2b2e42c7337055e1a15d5d5d501"} Dec 05 11:04:58.355849 master-0 kubenswrapper[24928]: I1205 11:04:58.355832 24928 scope.go:117] "RemoveContainer" containerID="06f90d83e4631f295c3fb423c8167bcab0f2e9c635020dbf8a539ff77f63ff64" Dec 05 11:04:58.355981 master-0 kubenswrapper[24928]: I1205 11:04:58.355880 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57545c5d5f-j2xrl" Dec 05 11:04:58.359126 master-0 kubenswrapper[24928]: I1205 11:04:58.359041 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vb5lw" event={"ID":"6592b5de-3923-472c-af99-f7ab0e1caf3f","Type":"ContainerDied","Data":"7e41a47bfc7ae80b9d9d0a990f3507e7d5934f474959949298c8f7a2296489a2"} Dec 05 11:04:58.359126 master-0 kubenswrapper[24928]: I1205 11:04:58.359084 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vb5lw" Dec 05 11:04:58.359126 master-0 kubenswrapper[24928]: I1205 11:04:58.359098 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e41a47bfc7ae80b9d9d0a990f3507e7d5934f474959949298c8f7a2296489a2" Dec 05 11:04:58.388693 master-0 kubenswrapper[24928]: I1205 11:04:58.388138 24928 scope.go:117] "RemoveContainer" containerID="2f405e7bb026d757ed7ec977a2c0bb77e68eae309598e289468fc835898db88c" Dec 05 11:04:59.005710 master-0 kubenswrapper[24928]: I1205 11:04:59.005650 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-tr2vj" Dec 05 11:04:59.010694 master-0 kubenswrapper[24928]: I1205 11:04:59.010589 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75b5-account-create-update-bpqzn" Dec 05 11:04:59.398272 master-0 kubenswrapper[24928]: I1205 11:04:59.398211 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-tr2vj" Dec 05 11:04:59.399194 master-0 kubenswrapper[24928]: I1205 11:04:59.398238 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-tr2vj" event={"ID":"c1ffd92e-85ff-4cab-a0c5-55968a23bc53","Type":"ContainerDied","Data":"4ca25df915105587da2490f94a3c11f9c9d27da7edb8087c492a8fd63e058999"} Dec 05 11:04:59.399410 master-0 kubenswrapper[24928]: I1205 11:04:59.399387 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ca25df915105587da2490f94a3c11f9c9d27da7edb8087c492a8fd63e058999" Dec 05 11:04:59.401883 master-0 kubenswrapper[24928]: I1205 11:04:59.401798 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-75b5-account-create-update-bpqzn" event={"ID":"ae071a88-af53-4fb3-93fc-76d8fd311de0","Type":"ContainerDied","Data":"5e37635c63537ba8ff70df6536db52e83e053e659d62485ce8041d7cec18525d"} Dec 05 11:04:59.402001 master-0 kubenswrapper[24928]: I1205 11:04:59.401896 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e37635c63537ba8ff70df6536db52e83e053e659d62485ce8041d7cec18525d" Dec 05 11:04:59.402001 master-0 kubenswrapper[24928]: I1205 11:04:59.401980 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-75b5-account-create-update-bpqzn" Dec 05 11:04:59.452746 master-0 kubenswrapper[24928]: I1205 11:04:59.452637 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49t8d\" (UniqueName: \"kubernetes.io/projected/ae071a88-af53-4fb3-93fc-76d8fd311de0-kube-api-access-49t8d\") pod \"ae071a88-af53-4fb3-93fc-76d8fd311de0\" (UID: \"ae071a88-af53-4fb3-93fc-76d8fd311de0\") " Dec 05 11:04:59.452974 master-0 kubenswrapper[24928]: I1205 11:04:59.452779 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghqzz\" (UniqueName: \"kubernetes.io/projected/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-kube-api-access-ghqzz\") pod \"c1ffd92e-85ff-4cab-a0c5-55968a23bc53\" (UID: \"c1ffd92e-85ff-4cab-a0c5-55968a23bc53\") " Dec 05 11:04:59.452974 master-0 kubenswrapper[24928]: I1205 11:04:59.452883 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae071a88-af53-4fb3-93fc-76d8fd311de0-operator-scripts\") pod \"ae071a88-af53-4fb3-93fc-76d8fd311de0\" (UID: \"ae071a88-af53-4fb3-93fc-76d8fd311de0\") " Dec 05 11:04:59.452974 master-0 kubenswrapper[24928]: I1205 11:04:59.452917 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-operator-scripts\") pod \"c1ffd92e-85ff-4cab-a0c5-55968a23bc53\" (UID: \"c1ffd92e-85ff-4cab-a0c5-55968a23bc53\") " Dec 05 11:04:59.453634 master-0 kubenswrapper[24928]: I1205 11:04:59.453576 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae071a88-af53-4fb3-93fc-76d8fd311de0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae071a88-af53-4fb3-93fc-76d8fd311de0" (UID: "ae071a88-af53-4fb3-93fc-76d8fd311de0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:59.453702 master-0 kubenswrapper[24928]: I1205 11:04:59.453664 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c1ffd92e-85ff-4cab-a0c5-55968a23bc53" (UID: "c1ffd92e-85ff-4cab-a0c5-55968a23bc53"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:04:59.454339 master-0 kubenswrapper[24928]: I1205 11:04:59.454308 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae071a88-af53-4fb3-93fc-76d8fd311de0-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:59.454397 master-0 kubenswrapper[24928]: I1205 11:04:59.454340 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:59.456539 master-0 kubenswrapper[24928]: I1205 11:04:59.456518 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-kube-api-access-ghqzz" (OuterVolumeSpecName: "kube-api-access-ghqzz") pod "c1ffd92e-85ff-4cab-a0c5-55968a23bc53" (UID: "c1ffd92e-85ff-4cab-a0c5-55968a23bc53"). InnerVolumeSpecName "kube-api-access-ghqzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:59.456751 master-0 kubenswrapper[24928]: I1205 11:04:59.456713 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae071a88-af53-4fb3-93fc-76d8fd311de0-kube-api-access-49t8d" (OuterVolumeSpecName: "kube-api-access-49t8d") pod "ae071a88-af53-4fb3-93fc-76d8fd311de0" (UID: "ae071a88-af53-4fb3-93fc-76d8fd311de0"). InnerVolumeSpecName "kube-api-access-49t8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:04:59.557564 master-0 kubenswrapper[24928]: I1205 11:04:59.557273 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49t8d\" (UniqueName: \"kubernetes.io/projected/ae071a88-af53-4fb3-93fc-76d8fd311de0-kube-api-access-49t8d\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:59.557564 master-0 kubenswrapper[24928]: I1205 11:04:59.557349 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghqzz\" (UniqueName: \"kubernetes.io/projected/c1ffd92e-85ff-4cab-a0c5-55968a23bc53-kube-api-access-ghqzz\") on node \"master-0\" DevicePath \"\"" Dec 05 11:04:59.884528 master-0 kubenswrapper[24928]: I1205 11:04:59.884406 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 05 11:05:00.042507 master-0 kubenswrapper[24928]: I1205 11:05:00.042433 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57545c5d5f-j2xrl"] Dec 05 11:05:00.061462 master-0 kubenswrapper[24928]: I1205 11:05:00.054201 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57545c5d5f-j2xrl"] Dec 05 11:05:00.141778 master-0 kubenswrapper[24928]: I1205 11:05:00.141703 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-bp44h"] Dec 05 11:05:00.142225 master-0 kubenswrapper[24928]: E1205 11:05:00.142193 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae071a88-af53-4fb3-93fc-76d8fd311de0" containerName="mariadb-account-create-update" Dec 05 11:05:00.142225 master-0 kubenswrapper[24928]: I1205 11:05:00.142220 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae071a88-af53-4fb3-93fc-76d8fd311de0" containerName="mariadb-account-create-update" Dec 05 11:05:00.142349 master-0 kubenswrapper[24928]: E1205 11:05:00.142233 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1ffd92e-85ff-4cab-a0c5-55968a23bc53" containerName="mariadb-database-create" Dec 05 11:05:00.142349 master-0 kubenswrapper[24928]: I1205 11:05:00.142241 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1ffd92e-85ff-4cab-a0c5-55968a23bc53" containerName="mariadb-database-create" Dec 05 11:05:00.142349 master-0 kubenswrapper[24928]: E1205 11:05:00.142267 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6592b5de-3923-472c-af99-f7ab0e1caf3f" containerName="mariadb-database-create" Dec 05 11:05:00.142349 master-0 kubenswrapper[24928]: I1205 11:05:00.142277 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6592b5de-3923-472c-af99-f7ab0e1caf3f" containerName="mariadb-database-create" Dec 05 11:05:00.142349 master-0 kubenswrapper[24928]: E1205 11:05:00.142293 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3249e2-8dd4-42e6-b78e-8712b33d7cf1" containerName="mariadb-account-create-update" Dec 05 11:05:00.142349 master-0 kubenswrapper[24928]: I1205 11:05:00.142302 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3249e2-8dd4-42e6-b78e-8712b33d7cf1" containerName="mariadb-account-create-update" Dec 05 11:05:00.142349 master-0 kubenswrapper[24928]: E1205 11:05:00.142327 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5337585b-1757-4276-99fc-672cf04a1a89" containerName="init" Dec 05 11:05:00.142349 master-0 kubenswrapper[24928]: I1205 11:05:00.142336 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5337585b-1757-4276-99fc-672cf04a1a89" containerName="init" Dec 05 11:05:00.142349 master-0 kubenswrapper[24928]: E1205 11:05:00.142351 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5337585b-1757-4276-99fc-672cf04a1a89" containerName="dnsmasq-dns" Dec 05 11:05:00.142764 master-0 kubenswrapper[24928]: I1205 11:05:00.142359 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5337585b-1757-4276-99fc-672cf04a1a89" containerName="dnsmasq-dns" Dec 05 11:05:00.142764 master-0 kubenswrapper[24928]: E1205 11:05:00.142373 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3af15ab-5fe0-4536-b61f-77c795f6e7f2" containerName="mariadb-database-create" Dec 05 11:05:00.142764 master-0 kubenswrapper[24928]: I1205 11:05:00.142382 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3af15ab-5fe0-4536-b61f-77c795f6e7f2" containerName="mariadb-database-create" Dec 05 11:05:00.142764 master-0 kubenswrapper[24928]: I1205 11:05:00.142715 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a3249e2-8dd4-42e6-b78e-8712b33d7cf1" containerName="mariadb-account-create-update" Dec 05 11:05:00.142764 master-0 kubenswrapper[24928]: I1205 11:05:00.142741 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5337585b-1757-4276-99fc-672cf04a1a89" containerName="dnsmasq-dns" Dec 05 11:05:00.142764 master-0 kubenswrapper[24928]: I1205 11:05:00.142764 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1ffd92e-85ff-4cab-a0c5-55968a23bc53" containerName="mariadb-database-create" Dec 05 11:05:00.142998 master-0 kubenswrapper[24928]: I1205 11:05:00.142792 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="6592b5de-3923-472c-af99-f7ab0e1caf3f" containerName="mariadb-database-create" Dec 05 11:05:00.142998 master-0 kubenswrapper[24928]: I1205 11:05:00.142809 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae071a88-af53-4fb3-93fc-76d8fd311de0" containerName="mariadb-account-create-update" Dec 05 11:05:00.142998 master-0 kubenswrapper[24928]: I1205 11:05:00.142828 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3af15ab-5fe0-4536-b61f-77c795f6e7f2" containerName="mariadb-database-create" Dec 05 11:05:00.143602 master-0 kubenswrapper[24928]: I1205 11:05:00.143570 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.148103 master-0 kubenswrapper[24928]: I1205 11:05:00.148013 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-ec941-config-data" Dec 05 11:05:00.165986 master-0 kubenswrapper[24928]: I1205 11:05:00.159144 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=57.971941838 podStartE2EDuration="1m1.159119011s" podCreationTimestamp="2025-12-05 11:03:59 +0000 UTC" firstStartedPulling="2025-12-05 11:04:19.235398849 +0000 UTC m=+1019.238592700" lastFinishedPulling="2025-12-05 11:04:22.422576022 +0000 UTC m=+1022.425769873" observedRunningTime="2025-12-05 11:05:00.153840769 +0000 UTC m=+1060.157034640" watchObservedRunningTime="2025-12-05 11:05:00.159119011 +0000 UTC m=+1060.162312872" Dec 05 11:05:00.173501 master-0 kubenswrapper[24928]: I1205 11:05:00.173401 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-combined-ca-bundle\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.173735 master-0 kubenswrapper[24928]: I1205 11:05:00.173520 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-db-sync-config-data\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.173735 master-0 kubenswrapper[24928]: I1205 11:05:00.173643 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdqcv\" (UniqueName: \"kubernetes.io/projected/1d10625d-a13a-4538-9d41-6e71f1af3669-kube-api-access-gdqcv\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.174073 master-0 kubenswrapper[24928]: I1205 11:05:00.173969 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-config-data\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.208730 master-0 kubenswrapper[24928]: I1205 11:05:00.208663 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-bp44h"] Dec 05 11:05:00.235194 master-0 kubenswrapper[24928]: I1205 11:05:00.235125 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5337585b-1757-4276-99fc-672cf04a1a89" path="/var/lib/kubelet/pods/5337585b-1757-4276-99fc-672cf04a1a89/volumes" Dec 05 11:05:00.281479 master-0 kubenswrapper[24928]: I1205 11:05:00.281354 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdqcv\" (UniqueName: \"kubernetes.io/projected/1d10625d-a13a-4538-9d41-6e71f1af3669-kube-api-access-gdqcv\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.281800 master-0 kubenswrapper[24928]: I1205 11:05:00.281759 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-config-data\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.282267 master-0 kubenswrapper[24928]: I1205 11:05:00.282052 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-combined-ca-bundle\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.282403 master-0 kubenswrapper[24928]: I1205 11:05:00.282274 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-db-sync-config-data\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.286792 master-0 kubenswrapper[24928]: I1205 11:05:00.286723 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-config-data\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.288141 master-0 kubenswrapper[24928]: I1205 11:05:00.288111 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-combined-ca-bundle\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.299201 master-0 kubenswrapper[24928]: I1205 11:05:00.299067 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-db-sync-config-data\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.308108 master-0 kubenswrapper[24928]: I1205 11:05:00.303955 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdqcv\" (UniqueName: \"kubernetes.io/projected/1d10625d-a13a-4538-9d41-6e71f1af3669-kube-api-access-gdqcv\") pod \"glance-db-sync-bp44h\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.350881 master-0 kubenswrapper[24928]: I1205 11:05:00.348043 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lvktj"] Dec 05 11:05:00.353642 master-0 kubenswrapper[24928]: I1205 11:05:00.353600 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:00.372607 master-0 kubenswrapper[24928]: I1205 11:05:00.364853 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lvktj"] Dec 05 11:05:00.386439 master-0 kubenswrapper[24928]: I1205 11:05:00.385156 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-utilities\") pod \"redhat-marketplace-lvktj\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:00.386439 master-0 kubenswrapper[24928]: I1205 11:05:00.385297 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-catalog-content\") pod \"redhat-marketplace-lvktj\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:00.386439 master-0 kubenswrapper[24928]: I1205 11:05:00.385449 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqwh7\" (UniqueName: \"kubernetes.io/projected/0e07646a-fd0f-49fc-89da-5938c26aad18-kube-api-access-pqwh7\") pod \"redhat-marketplace-lvktj\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:00.484046 master-0 kubenswrapper[24928]: I1205 11:05:00.483954 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:00.487355 master-0 kubenswrapper[24928]: I1205 11:05:00.487199 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-utilities\") pod \"redhat-marketplace-lvktj\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:00.487651 master-0 kubenswrapper[24928]: I1205 11:05:00.487399 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-catalog-content\") pod \"redhat-marketplace-lvktj\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:00.487651 master-0 kubenswrapper[24928]: I1205 11:05:00.487491 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqwh7\" (UniqueName: \"kubernetes.io/projected/0e07646a-fd0f-49fc-89da-5938c26aad18-kube-api-access-pqwh7\") pod \"redhat-marketplace-lvktj\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:00.487874 master-0 kubenswrapper[24928]: I1205 11:05:00.487824 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-utilities\") pod \"redhat-marketplace-lvktj\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:00.488327 master-0 kubenswrapper[24928]: I1205 11:05:00.488281 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-catalog-content\") pod \"redhat-marketplace-lvktj\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:00.508668 master-0 kubenswrapper[24928]: I1205 11:05:00.507566 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqwh7\" (UniqueName: \"kubernetes.io/projected/0e07646a-fd0f-49fc-89da-5938c26aad18-kube-api-access-pqwh7\") pod \"redhat-marketplace-lvktj\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:00.724758 master-0 kubenswrapper[24928]: I1205 11:05:00.724047 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:01.214510 master-0 kubenswrapper[24928]: I1205 11:05:01.214461 24928 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:05:01.219272 master-0 kubenswrapper[24928]: I1205 11:05:01.219198 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-bp44h"] Dec 05 11:05:01.354448 master-0 kubenswrapper[24928]: I1205 11:05:01.354335 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lvktj"] Dec 05 11:05:01.355976 master-0 kubenswrapper[24928]: W1205 11:05:01.355918 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e07646a_fd0f_49fc_89da_5938c26aad18.slice/crio-97ed2f6a5693ad410a9131bea71f95a34435534fae8d925fcb7524ccba134f62 WatchSource:0}: Error finding container 97ed2f6a5693ad410a9131bea71f95a34435534fae8d925fcb7524ccba134f62: Status 404 returned error can't find the container with id 97ed2f6a5693ad410a9131bea71f95a34435534fae8d925fcb7524ccba134f62 Dec 05 11:05:01.436486 master-0 kubenswrapper[24928]: I1205 11:05:01.436401 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvktj" event={"ID":"0e07646a-fd0f-49fc-89da-5938c26aad18","Type":"ContainerStarted","Data":"97ed2f6a5693ad410a9131bea71f95a34435534fae8d925fcb7524ccba134f62"} Dec 05 11:05:01.437836 master-0 kubenswrapper[24928]: I1205 11:05:01.437710 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bp44h" event={"ID":"1d10625d-a13a-4538-9d41-6e71f1af3669","Type":"ContainerStarted","Data":"c83fefe62f19dca5c51029a78a321b7f583eb37db5813c2105addbd16ee31f81"} Dec 05 11:05:02.449726 master-0 kubenswrapper[24928]: I1205 11:05:02.449634 24928 generic.go:334] "Generic (PLEG): container finished" podID="0e07646a-fd0f-49fc-89da-5938c26aad18" containerID="e7546ce87acba6b8e0104bd03704f38f94b6ae6c375b852b6cc252028b5419ab" exitCode=0 Dec 05 11:05:02.451441 master-0 kubenswrapper[24928]: I1205 11:05:02.449705 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvktj" event={"ID":"0e07646a-fd0f-49fc-89da-5938c26aad18","Type":"ContainerDied","Data":"e7546ce87acba6b8e0104bd03704f38f94b6ae6c375b852b6cc252028b5419ab"} Dec 05 11:05:03.465000 master-0 kubenswrapper[24928]: I1205 11:05:03.464843 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvktj" event={"ID":"0e07646a-fd0f-49fc-89da-5938c26aad18","Type":"ContainerStarted","Data":"d66ef34a556dbec95523b1b94b6f318111a67e1318d2494a1fff969cd14b67dc"} Dec 05 11:05:03.469183 master-0 kubenswrapper[24928]: I1205 11:05:03.469125 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b2rmc" event={"ID":"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28","Type":"ContainerDied","Data":"4b3ebf5f20bc7fd19b8fd2c7e9c45f05dc8fd24c06c885d0dafccdbf308eb008"} Dec 05 11:05:03.469183 master-0 kubenswrapper[24928]: I1205 11:05:03.469159 24928 generic.go:334] "Generic (PLEG): container finished" podID="61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28" containerID="4b3ebf5f20bc7fd19b8fd2c7e9c45f05dc8fd24c06c885d0dafccdbf308eb008" exitCode=0 Dec 05 11:05:04.275015 master-0 kubenswrapper[24928]: I1205 11:05:04.274939 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:05:04.279141 master-0 kubenswrapper[24928]: I1205 11:05:04.279101 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cc6a7a43-f73c-4e31-a61e-de8f4a5957b6-etc-swift\") pod \"swift-storage-0\" (UID: \"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6\") " pod="openstack/swift-storage-0" Dec 05 11:05:04.405078 master-0 kubenswrapper[24928]: I1205 11:05:04.384052 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 05 11:05:04.523171 master-0 kubenswrapper[24928]: I1205 11:05:04.523090 24928 generic.go:334] "Generic (PLEG): container finished" podID="0e07646a-fd0f-49fc-89da-5938c26aad18" containerID="d66ef34a556dbec95523b1b94b6f318111a67e1318d2494a1fff969cd14b67dc" exitCode=0 Dec 05 11:05:04.523672 master-0 kubenswrapper[24928]: I1205 11:05:04.523243 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvktj" event={"ID":"0e07646a-fd0f-49fc-89da-5938c26aad18","Type":"ContainerDied","Data":"d66ef34a556dbec95523b1b94b6f318111a67e1318d2494a1fff969cd14b67dc"} Dec 05 11:05:04.709257 master-0 kubenswrapper[24928]: I1205 11:05:04.708976 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-rz85b" podUID="24ba60fc-8a32-4d40-973f-ccabca97d566" containerName="ovn-controller" probeResult="failure" output=< Dec 05 11:05:04.709257 master-0 kubenswrapper[24928]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 05 11:05:04.709257 master-0 kubenswrapper[24928]: > Dec 05 11:05:04.822137 master-0 kubenswrapper[24928]: I1205 11:05:04.822091 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:05:04.831029 master-0 kubenswrapper[24928]: I1205 11:05:04.831004 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-rvm7b" Dec 05 11:05:05.109921 master-0 kubenswrapper[24928]: I1205 11:05:05.109797 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:05:05.274329 master-0 kubenswrapper[24928]: I1205 11:05:05.274216 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-ring-data-devices\") pod \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " Dec 05 11:05:05.274570 master-0 kubenswrapper[24928]: I1205 11:05:05.274414 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-etc-swift\") pod \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " Dec 05 11:05:05.274570 master-0 kubenswrapper[24928]: I1205 11:05:05.274551 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75vzp\" (UniqueName: \"kubernetes.io/projected/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-kube-api-access-75vzp\") pod \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " Dec 05 11:05:05.274701 master-0 kubenswrapper[24928]: I1205 11:05:05.274613 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-dispersionconf\") pod \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " Dec 05 11:05:05.274760 master-0 kubenswrapper[24928]: I1205 11:05:05.274735 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-scripts\") pod \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " Dec 05 11:05:05.275038 master-0 kubenswrapper[24928]: I1205 11:05:05.274974 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28" (UID: "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:05.275117 master-0 kubenswrapper[24928]: I1205 11:05:05.274977 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-combined-ca-bundle\") pod \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " Dec 05 11:05:05.275172 master-0 kubenswrapper[24928]: I1205 11:05:05.275152 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-swiftconf\") pod \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\" (UID: \"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28\") " Dec 05 11:05:05.275613 master-0 kubenswrapper[24928]: I1205 11:05:05.275348 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28" (UID: "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:05:05.276786 master-0 kubenswrapper[24928]: I1205 11:05:05.276742 24928 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-ring-data-devices\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:05.276786 master-0 kubenswrapper[24928]: I1205 11:05:05.276780 24928 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-etc-swift\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:05.278265 master-0 kubenswrapper[24928]: I1205 11:05:05.278210 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-kube-api-access-75vzp" (OuterVolumeSpecName: "kube-api-access-75vzp") pod "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28" (UID: "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28"). InnerVolumeSpecName "kube-api-access-75vzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:05.280301 master-0 kubenswrapper[24928]: I1205 11:05:05.280252 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28" (UID: "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:05.301254 master-0 kubenswrapper[24928]: I1205 11:05:05.301148 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-scripts" (OuterVolumeSpecName: "scripts") pod "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28" (UID: "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:05.320366 master-0 kubenswrapper[24928]: I1205 11:05:05.320263 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28" (UID: "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:05.332397 master-0 kubenswrapper[24928]: I1205 11:05:05.332086 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28" (UID: "61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:05.382613 master-0 kubenswrapper[24928]: I1205 11:05:05.378796 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75vzp\" (UniqueName: \"kubernetes.io/projected/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-kube-api-access-75vzp\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:05.382613 master-0 kubenswrapper[24928]: I1205 11:05:05.378847 24928 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-dispersionconf\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:05.382613 master-0 kubenswrapper[24928]: I1205 11:05:05.378867 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:05.382613 master-0 kubenswrapper[24928]: I1205 11:05:05.378878 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:05.382613 master-0 kubenswrapper[24928]: I1205 11:05:05.378888 24928 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28-swiftconf\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:05.536514 master-0 kubenswrapper[24928]: I1205 11:05:05.536447 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-b2rmc" Dec 05 11:05:05.536993 master-0 kubenswrapper[24928]: I1205 11:05:05.536730 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-b2rmc" event={"ID":"61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28","Type":"ContainerDied","Data":"93f80fd6da9673e6e1738d7eda3dcecf4824563cf4a111d3d6e341f19afcfcd5"} Dec 05 11:05:05.536993 master-0 kubenswrapper[24928]: I1205 11:05:05.536786 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93f80fd6da9673e6e1738d7eda3dcecf4824563cf4a111d3d6e341f19afcfcd5" Dec 05 11:05:05.756456 master-0 kubenswrapper[24928]: I1205 11:05:05.753983 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 05 11:05:05.915343 master-0 kubenswrapper[24928]: I1205 11:05:05.909216 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-rz85b-config-cxcz8"] Dec 05 11:05:05.915343 master-0 kubenswrapper[24928]: E1205 11:05:05.915028 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28" containerName="swift-ring-rebalance" Dec 05 11:05:05.915343 master-0 kubenswrapper[24928]: I1205 11:05:05.915083 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28" containerName="swift-ring-rebalance" Dec 05 11:05:05.915719 master-0 kubenswrapper[24928]: I1205 11:05:05.915380 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="61ea5bac-ea53-4d1e-aaea-2b1ad9a0ca28" containerName="swift-ring-rebalance" Dec 05 11:05:05.921274 master-0 kubenswrapper[24928]: I1205 11:05:05.916332 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:05.921274 master-0 kubenswrapper[24928]: I1205 11:05:05.921183 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 05 11:05:05.942462 master-0 kubenswrapper[24928]: I1205 11:05:05.938601 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rz85b-config-cxcz8"] Dec 05 11:05:06.016215 master-0 kubenswrapper[24928]: I1205 11:05:06.007375 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-log-ovn\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.016215 master-0 kubenswrapper[24928]: I1205 11:05:06.007443 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.016215 master-0 kubenswrapper[24928]: I1205 11:05:06.007506 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run-ovn\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.016215 master-0 kubenswrapper[24928]: I1205 11:05:06.007530 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxxpf\" (UniqueName: \"kubernetes.io/projected/8fe5c439-67b4-4961-a197-736e934f5950-kube-api-access-vxxpf\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.016215 master-0 kubenswrapper[24928]: I1205 11:05:06.007565 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-additional-scripts\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.016215 master-0 kubenswrapper[24928]: I1205 11:05:06.007592 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-scripts\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.109939 master-0 kubenswrapper[24928]: I1205 11:05:06.109102 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-log-ovn\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.109939 master-0 kubenswrapper[24928]: I1205 11:05:06.109158 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.109939 master-0 kubenswrapper[24928]: I1205 11:05:06.109315 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run-ovn\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.109939 master-0 kubenswrapper[24928]: I1205 11:05:06.109344 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxxpf\" (UniqueName: \"kubernetes.io/projected/8fe5c439-67b4-4961-a197-736e934f5950-kube-api-access-vxxpf\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.109939 master-0 kubenswrapper[24928]: I1205 11:05:06.109380 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-additional-scripts\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.109939 master-0 kubenswrapper[24928]: I1205 11:05:06.109398 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-log-ovn\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.109939 master-0 kubenswrapper[24928]: I1205 11:05:06.109436 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-scripts\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.110396 master-0 kubenswrapper[24928]: I1205 11:05:06.110352 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run-ovn\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.110557 master-0 kubenswrapper[24928]: I1205 11:05:06.110534 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.111712 master-0 kubenswrapper[24928]: I1205 11:05:06.111688 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-scripts\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.111804 master-0 kubenswrapper[24928]: I1205 11:05:06.111741 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-additional-scripts\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.135479 master-0 kubenswrapper[24928]: I1205 11:05:06.135193 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxxpf\" (UniqueName: \"kubernetes.io/projected/8fe5c439-67b4-4961-a197-736e934f5950-kube-api-access-vxxpf\") pod \"ovn-controller-rz85b-config-cxcz8\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.391998 master-0 kubenswrapper[24928]: I1205 11:05:06.391940 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:06.549672 master-0 kubenswrapper[24928]: I1205 11:05:06.549581 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"65cacc65bdf8cc746918d2db7bdb3a60e8dff3b012a9c2760236e30be23ff587"} Dec 05 11:05:06.552138 master-0 kubenswrapper[24928]: I1205 11:05:06.551773 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvktj" event={"ID":"0e07646a-fd0f-49fc-89da-5938c26aad18","Type":"ContainerStarted","Data":"9cf3b503a6c68397953696990b7e6646d8f3a7dd04827ef76e79e29506ae342a"} Dec 05 11:05:07.101686 master-0 kubenswrapper[24928]: I1205 11:05:07.101610 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 05 11:05:09.004590 master-0 kubenswrapper[24928]: I1205 11:05:09.004496 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-rz85b-config-cxcz8"] Dec 05 11:05:09.012649 master-0 kubenswrapper[24928]: I1205 11:05:09.012527 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lvktj" podStartSLOduration=5.752075005 podStartE2EDuration="9.012508244s" podCreationTimestamp="2025-12-05 11:05:00 +0000 UTC" firstStartedPulling="2025-12-05 11:05:02.45290032 +0000 UTC m=+1062.456094161" lastFinishedPulling="2025-12-05 11:05:05.713333549 +0000 UTC m=+1065.716527400" observedRunningTime="2025-12-05 11:05:08.999217333 +0000 UTC m=+1069.002411184" watchObservedRunningTime="2025-12-05 11:05:09.012508244 +0000 UTC m=+1069.015702115" Dec 05 11:05:09.605050 master-0 kubenswrapper[24928]: I1205 11:05:09.604903 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rz85b-config-cxcz8" event={"ID":"8fe5c439-67b4-4961-a197-736e934f5950","Type":"ContainerStarted","Data":"8b366a44377a553202afffd5e196805a86d97c265f2d4791b8be0632a6bc7543"} Dec 05 11:05:09.605050 master-0 kubenswrapper[24928]: I1205 11:05:09.604989 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rz85b-config-cxcz8" event={"ID":"8fe5c439-67b4-4961-a197-736e934f5950","Type":"ContainerStarted","Data":"9a71be9063742dfc9430d0365f2e40fd7ef13de01c7075f7ca89fd1af493abc0"} Dec 05 11:05:09.680756 master-0 kubenswrapper[24928]: I1205 11:05:09.680681 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-rz85b" Dec 05 11:05:10.725586 master-0 kubenswrapper[24928]: I1205 11:05:10.725484 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:10.725586 master-0 kubenswrapper[24928]: I1205 11:05:10.725573 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:10.791919 master-0 kubenswrapper[24928]: I1205 11:05:10.791836 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:11.680228 master-0 kubenswrapper[24928]: I1205 11:05:11.680114 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:15.650965 master-0 kubenswrapper[24928]: I1205 11:05:15.650794 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 05 11:05:17.536928 master-0 kubenswrapper[24928]: I1205 11:05:17.536820 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-rz85b-config-cxcz8" podStartSLOduration=12.536793019 podStartE2EDuration="12.536793019s" podCreationTimestamp="2025-12-05 11:05:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:05:17.53322718 +0000 UTC m=+1077.536421041" watchObservedRunningTime="2025-12-05 11:05:17.536793019 +0000 UTC m=+1077.539986870" Dec 05 11:05:17.703620 master-0 kubenswrapper[24928]: I1205 11:05:17.702500 24928 generic.go:334] "Generic (PLEG): container finished" podID="8fe5c439-67b4-4961-a197-736e934f5950" containerID="8b366a44377a553202afffd5e196805a86d97c265f2d4791b8be0632a6bc7543" exitCode=0 Dec 05 11:05:17.703620 master-0 kubenswrapper[24928]: I1205 11:05:17.702610 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rz85b-config-cxcz8" event={"ID":"8fe5c439-67b4-4961-a197-736e934f5950","Type":"ContainerDied","Data":"8b366a44377a553202afffd5e196805a86d97c265f2d4791b8be0632a6bc7543"} Dec 05 11:05:18.964254 master-0 kubenswrapper[24928]: I1205 11:05:18.962231 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vdmr2"] Dec 05 11:05:18.967430 master-0 kubenswrapper[24928]: I1205 11:05:18.966381 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:18.983380 master-0 kubenswrapper[24928]: I1205 11:05:18.983122 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x5nq4"] Dec 05 11:05:18.994705 master-0 kubenswrapper[24928]: I1205 11:05:18.994599 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vdmr2"] Dec 05 11:05:18.994705 master-0 kubenswrapper[24928]: I1205 11:05:18.994706 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:19.055472 master-0 kubenswrapper[24928]: I1205 11:05:19.046064 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x5nq4"] Dec 05 11:05:19.115634 master-0 kubenswrapper[24928]: I1205 11:05:19.115378 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfj6w\" (UniqueName: \"kubernetes.io/projected/98a4e639-cc2c-406e-b1b7-48746557cdc4-kube-api-access-cfj6w\") pod \"redhat-operators-vdmr2\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:19.115870 master-0 kubenswrapper[24928]: I1205 11:05:19.115806 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-catalog-content\") pod \"redhat-operators-vdmr2\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:19.115924 master-0 kubenswrapper[24928]: I1205 11:05:19.115904 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9z8g\" (UniqueName: \"kubernetes.io/projected/032f0939-20e7-46cb-9665-b1346b7e1b10-kube-api-access-x9z8g\") pod \"certified-operators-x5nq4\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:19.116108 master-0 kubenswrapper[24928]: I1205 11:05:19.116081 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-utilities\") pod \"redhat-operators-vdmr2\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:19.116161 master-0 kubenswrapper[24928]: I1205 11:05:19.116144 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-utilities\") pod \"certified-operators-x5nq4\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:19.116695 master-0 kubenswrapper[24928]: I1205 11:05:19.116497 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-catalog-content\") pod \"certified-operators-x5nq4\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:19.220785 master-0 kubenswrapper[24928]: I1205 11:05:19.220722 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfj6w\" (UniqueName: \"kubernetes.io/projected/98a4e639-cc2c-406e-b1b7-48746557cdc4-kube-api-access-cfj6w\") pod \"redhat-operators-vdmr2\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:19.221017 master-0 kubenswrapper[24928]: I1205 11:05:19.220860 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-catalog-content\") pod \"redhat-operators-vdmr2\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:19.221017 master-0 kubenswrapper[24928]: I1205 11:05:19.220891 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9z8g\" (UniqueName: \"kubernetes.io/projected/032f0939-20e7-46cb-9665-b1346b7e1b10-kube-api-access-x9z8g\") pod \"certified-operators-x5nq4\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:19.221017 master-0 kubenswrapper[24928]: I1205 11:05:19.220951 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-utilities\") pod \"redhat-operators-vdmr2\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:19.221017 master-0 kubenswrapper[24928]: I1205 11:05:19.220983 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-utilities\") pod \"certified-operators-x5nq4\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:19.221215 master-0 kubenswrapper[24928]: I1205 11:05:19.221044 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-catalog-content\") pod \"certified-operators-x5nq4\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:19.221665 master-0 kubenswrapper[24928]: I1205 11:05:19.221633 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-catalog-content\") pod \"certified-operators-x5nq4\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:19.233653 master-0 kubenswrapper[24928]: I1205 11:05:19.232371 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-catalog-content\") pod \"redhat-operators-vdmr2\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:19.233653 master-0 kubenswrapper[24928]: I1205 11:05:19.232527 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-utilities\") pod \"certified-operators-x5nq4\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:19.233653 master-0 kubenswrapper[24928]: I1205 11:05:19.232568 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-utilities\") pod \"redhat-operators-vdmr2\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:19.662937 master-0 kubenswrapper[24928]: I1205 11:05:19.662862 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfj6w\" (UniqueName: \"kubernetes.io/projected/98a4e639-cc2c-406e-b1b7-48746557cdc4-kube-api-access-cfj6w\") pod \"redhat-operators-vdmr2\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:19.664659 master-0 kubenswrapper[24928]: I1205 11:05:19.664618 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9z8g\" (UniqueName: \"kubernetes.io/projected/032f0939-20e7-46cb-9665-b1346b7e1b10-kube-api-access-x9z8g\") pod \"certified-operators-x5nq4\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:19.756024 master-0 kubenswrapper[24928]: I1205 11:05:19.755953 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-rz85b-config-cxcz8" event={"ID":"8fe5c439-67b4-4961-a197-736e934f5950","Type":"ContainerDied","Data":"9a71be9063742dfc9430d0365f2e40fd7ef13de01c7075f7ca89fd1af493abc0"} Dec 05 11:05:19.756024 master-0 kubenswrapper[24928]: I1205 11:05:19.756011 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a71be9063742dfc9430d0365f2e40fd7ef13de01c7075f7ca89fd1af493abc0" Dec 05 11:05:19.757578 master-0 kubenswrapper[24928]: I1205 11:05:19.757548 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:19.853109 master-0 kubenswrapper[24928]: I1205 11:05:19.853039 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run-ovn\") pod \"8fe5c439-67b4-4961-a197-736e934f5950\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " Dec 05 11:05:19.853360 master-0 kubenswrapper[24928]: I1205 11:05:19.853132 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-additional-scripts\") pod \"8fe5c439-67b4-4961-a197-736e934f5950\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " Dec 05 11:05:19.853360 master-0 kubenswrapper[24928]: I1205 11:05:19.853180 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-scripts\") pod \"8fe5c439-67b4-4961-a197-736e934f5950\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " Dec 05 11:05:19.853360 master-0 kubenswrapper[24928]: I1205 11:05:19.853294 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxxpf\" (UniqueName: \"kubernetes.io/projected/8fe5c439-67b4-4961-a197-736e934f5950-kube-api-access-vxxpf\") pod \"8fe5c439-67b4-4961-a197-736e934f5950\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " Dec 05 11:05:19.853500 master-0 kubenswrapper[24928]: I1205 11:05:19.853450 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run\") pod \"8fe5c439-67b4-4961-a197-736e934f5950\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " Dec 05 11:05:19.853606 master-0 kubenswrapper[24928]: I1205 11:05:19.853578 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-log-ovn\") pod \"8fe5c439-67b4-4961-a197-736e934f5950\" (UID: \"8fe5c439-67b4-4961-a197-736e934f5950\") " Dec 05 11:05:19.857452 master-0 kubenswrapper[24928]: I1205 11:05:19.854197 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8fe5c439-67b4-4961-a197-736e934f5950" (UID: "8fe5c439-67b4-4961-a197-736e934f5950"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:05:19.857452 master-0 kubenswrapper[24928]: I1205 11:05:19.854318 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run" (OuterVolumeSpecName: "var-run") pod "8fe5c439-67b4-4961-a197-736e934f5950" (UID: "8fe5c439-67b4-4961-a197-736e934f5950"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:05:19.857452 master-0 kubenswrapper[24928]: I1205 11:05:19.854347 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8fe5c439-67b4-4961-a197-736e934f5950" (UID: "8fe5c439-67b4-4961-a197-736e934f5950"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:05:19.857452 master-0 kubenswrapper[24928]: I1205 11:05:19.855193 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-scripts" (OuterVolumeSpecName: "scripts") pod "8fe5c439-67b4-4961-a197-736e934f5950" (UID: "8fe5c439-67b4-4961-a197-736e934f5950"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:19.861444 master-0 kubenswrapper[24928]: I1205 11:05:19.858229 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "8fe5c439-67b4-4961-a197-736e934f5950" (UID: "8fe5c439-67b4-4961-a197-736e934f5950"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:19.874472 master-0 kubenswrapper[24928]: I1205 11:05:19.873358 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fe5c439-67b4-4961-a197-736e934f5950-kube-api-access-vxxpf" (OuterVolumeSpecName: "kube-api-access-vxxpf") pod "8fe5c439-67b4-4961-a197-736e934f5950" (UID: "8fe5c439-67b4-4961-a197-736e934f5950"). InnerVolumeSpecName "kube-api-access-vxxpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:19.921562 master-0 kubenswrapper[24928]: I1205 11:05:19.919557 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:19.934106 master-0 kubenswrapper[24928]: I1205 11:05:19.933994 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:19.968284 master-0 kubenswrapper[24928]: I1205 11:05:19.968211 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxxpf\" (UniqueName: \"kubernetes.io/projected/8fe5c439-67b4-4961-a197-736e934f5950-kube-api-access-vxxpf\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:19.968284 master-0 kubenswrapper[24928]: I1205 11:05:19.968260 24928 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:19.968284 master-0 kubenswrapper[24928]: I1205 11:05:19.968270 24928 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-log-ovn\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:19.968284 master-0 kubenswrapper[24928]: I1205 11:05:19.968278 24928 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8fe5c439-67b4-4961-a197-736e934f5950-var-run-ovn\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:19.968284 master-0 kubenswrapper[24928]: I1205 11:05:19.968287 24928 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-additional-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:19.968284 master-0 kubenswrapper[24928]: I1205 11:05:19.968301 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8fe5c439-67b4-4961-a197-736e934f5950-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:20.580465 master-0 kubenswrapper[24928]: I1205 11:05:20.576159 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x5nq4"] Dec 05 11:05:20.609465 master-0 kubenswrapper[24928]: W1205 11:05:20.607532 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod032f0939_20e7_46cb_9665_b1346b7e1b10.slice/crio-234dc119535ca140563a43f2ddcb668024e8e834bb3944505096db4a7b86e233 WatchSource:0}: Error finding container 234dc119535ca140563a43f2ddcb668024e8e834bb3944505096db4a7b86e233: Status 404 returned error can't find the container with id 234dc119535ca140563a43f2ddcb668024e8e834bb3944505096db4a7b86e233 Dec 05 11:05:20.678461 master-0 kubenswrapper[24928]: I1205 11:05:20.675129 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vdmr2"] Dec 05 11:05:20.695658 master-0 kubenswrapper[24928]: W1205 11:05:20.694574 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98a4e639_cc2c_406e_b1b7_48746557cdc4.slice/crio-da7afeabaace30849cec2c1dd38c2bdddf8fc79ff13b68dc39be74a112a8a416 WatchSource:0}: Error finding container da7afeabaace30849cec2c1dd38c2bdddf8fc79ff13b68dc39be74a112a8a416: Status 404 returned error can't find the container with id da7afeabaace30849cec2c1dd38c2bdddf8fc79ff13b68dc39be74a112a8a416 Dec 05 11:05:20.820713 master-0 kubenswrapper[24928]: I1205 11:05:20.820041 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdmr2" event={"ID":"98a4e639-cc2c-406e-b1b7-48746557cdc4","Type":"ContainerStarted","Data":"da7afeabaace30849cec2c1dd38c2bdddf8fc79ff13b68dc39be74a112a8a416"} Dec 05 11:05:20.826485 master-0 kubenswrapper[24928]: I1205 11:05:20.824436 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5nq4" event={"ID":"032f0939-20e7-46cb-9665-b1346b7e1b10","Type":"ContainerStarted","Data":"234dc119535ca140563a43f2ddcb668024e8e834bb3944505096db4a7b86e233"} Dec 05 11:05:20.826962 master-0 kubenswrapper[24928]: I1205 11:05:20.826925 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"5bb402707d18de022ba61839915dd0f470019988807e870b1342032c3aa866b7"} Dec 05 11:05:20.827035 master-0 kubenswrapper[24928]: I1205 11:05:20.826967 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"32b14685aa1928f5c93d285f2242de68043b52aca5dc42d8ab611a852b71e13b"} Dec 05 11:05:20.827035 master-0 kubenswrapper[24928]: I1205 11:05:20.826981 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"6a6fca3c7356ff19de5b8c73a973ec5eebbf269493be76f54203bd635ec9ef08"} Dec 05 11:05:20.838355 master-0 kubenswrapper[24928]: I1205 11:05:20.836619 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-rz85b-config-cxcz8" Dec 05 11:05:20.838355 master-0 kubenswrapper[24928]: I1205 11:05:20.836845 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bp44h" event={"ID":"1d10625d-a13a-4538-9d41-6e71f1af3669","Type":"ContainerStarted","Data":"3370f573d17f498b4ea25781e741647ab3a6193bb5b0f8a6e64abaa22c3dab7d"} Dec 05 11:05:20.938454 master-0 kubenswrapper[24928]: I1205 11:05:20.936662 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-bp44h" podStartSLOduration=2.3330898429999998 podStartE2EDuration="20.936639055s" podCreationTimestamp="2025-12-05 11:05:00 +0000 UTC" firstStartedPulling="2025-12-05 11:05:01.21435357 +0000 UTC m=+1061.217547411" lastFinishedPulling="2025-12-05 11:05:19.817902772 +0000 UTC m=+1079.821096623" observedRunningTime="2025-12-05 11:05:20.879785126 +0000 UTC m=+1080.882978977" watchObservedRunningTime="2025-12-05 11:05:20.936639055 +0000 UTC m=+1080.939832906" Dec 05 11:05:21.001617 master-0 kubenswrapper[24928]: I1205 11:05:21.001488 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-kf7cj"] Dec 05 11:05:21.002223 master-0 kubenswrapper[24928]: E1205 11:05:21.001979 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe5c439-67b4-4961-a197-736e934f5950" containerName="ovn-config" Dec 05 11:05:21.002223 master-0 kubenswrapper[24928]: I1205 11:05:21.002000 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe5c439-67b4-4961-a197-736e934f5950" containerName="ovn-config" Dec 05 11:05:21.002328 master-0 kubenswrapper[24928]: I1205 11:05:21.002275 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe5c439-67b4-4961-a197-736e934f5950" containerName="ovn-config" Dec 05 11:05:21.003379 master-0 kubenswrapper[24928]: I1205 11:05:21.003059 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kf7cj" Dec 05 11:05:21.038175 master-0 kubenswrapper[24928]: I1205 11:05:21.038061 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-kf7cj"] Dec 05 11:05:21.101465 master-0 kubenswrapper[24928]: I1205 11:05:21.099122 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-rz85b-config-cxcz8"] Dec 05 11:05:21.110293 master-0 kubenswrapper[24928]: I1205 11:05:21.110217 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtqqt\" (UniqueName: \"kubernetes.io/projected/17f4e63e-29f0-4884-a20e-a3cb02c85235-kube-api-access-wtqqt\") pod \"cinder-db-create-kf7cj\" (UID: \"17f4e63e-29f0-4884-a20e-a3cb02c85235\") " pod="openstack/cinder-db-create-kf7cj" Dec 05 11:05:21.110525 master-0 kubenswrapper[24928]: I1205 11:05:21.110450 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17f4e63e-29f0-4884-a20e-a3cb02c85235-operator-scripts\") pod \"cinder-db-create-kf7cj\" (UID: \"17f4e63e-29f0-4884-a20e-a3cb02c85235\") " pod="openstack/cinder-db-create-kf7cj" Dec 05 11:05:21.148550 master-0 kubenswrapper[24928]: I1205 11:05:21.148498 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-rz85b-config-cxcz8"] Dec 05 11:05:21.171594 master-0 kubenswrapper[24928]: I1205 11:05:21.167680 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-72f7-account-create-update-rb6rq"] Dec 05 11:05:21.171594 master-0 kubenswrapper[24928]: I1205 11:05:21.169546 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72f7-account-create-update-rb6rq" Dec 05 11:05:21.182209 master-0 kubenswrapper[24928]: I1205 11:05:21.174869 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 05 11:05:21.182209 master-0 kubenswrapper[24928]: I1205 11:05:21.179632 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-72f7-account-create-update-rb6rq"] Dec 05 11:05:21.215520 master-0 kubenswrapper[24928]: I1205 11:05:21.213891 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17f4e63e-29f0-4884-a20e-a3cb02c85235-operator-scripts\") pod \"cinder-db-create-kf7cj\" (UID: \"17f4e63e-29f0-4884-a20e-a3cb02c85235\") " pod="openstack/cinder-db-create-kf7cj" Dec 05 11:05:21.215520 master-0 kubenswrapper[24928]: I1205 11:05:21.213979 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtqqt\" (UniqueName: \"kubernetes.io/projected/17f4e63e-29f0-4884-a20e-a3cb02c85235-kube-api-access-wtqqt\") pod \"cinder-db-create-kf7cj\" (UID: \"17f4e63e-29f0-4884-a20e-a3cb02c85235\") " pod="openstack/cinder-db-create-kf7cj" Dec 05 11:05:21.215520 master-0 kubenswrapper[24928]: I1205 11:05:21.215386 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17f4e63e-29f0-4884-a20e-a3cb02c85235-operator-scripts\") pod \"cinder-db-create-kf7cj\" (UID: \"17f4e63e-29f0-4884-a20e-a3cb02c85235\") " pod="openstack/cinder-db-create-kf7cj" Dec 05 11:05:21.251363 master-0 kubenswrapper[24928]: I1205 11:05:21.251241 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtqqt\" (UniqueName: \"kubernetes.io/projected/17f4e63e-29f0-4884-a20e-a3cb02c85235-kube-api-access-wtqqt\") pod \"cinder-db-create-kf7cj\" (UID: \"17f4e63e-29f0-4884-a20e-a3cb02c85235\") " pod="openstack/cinder-db-create-kf7cj" Dec 05 11:05:21.318595 master-0 kubenswrapper[24928]: I1205 11:05:21.315927 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7wp8\" (UniqueName: \"kubernetes.io/projected/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-kube-api-access-v7wp8\") pod \"cinder-72f7-account-create-update-rb6rq\" (UID: \"1c3b42bc-736b-4910-9bef-39ebeba0c5dd\") " pod="openstack/cinder-72f7-account-create-update-rb6rq" Dec 05 11:05:21.318595 master-0 kubenswrapper[24928]: I1205 11:05:21.316128 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-operator-scripts\") pod \"cinder-72f7-account-create-update-rb6rq\" (UID: \"1c3b42bc-736b-4910-9bef-39ebeba0c5dd\") " pod="openstack/cinder-72f7-account-create-update-rb6rq" Dec 05 11:05:21.421628 master-0 kubenswrapper[24928]: I1205 11:05:21.417903 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-operator-scripts\") pod \"cinder-72f7-account-create-update-rb6rq\" (UID: \"1c3b42bc-736b-4910-9bef-39ebeba0c5dd\") " pod="openstack/cinder-72f7-account-create-update-rb6rq" Dec 05 11:05:21.421628 master-0 kubenswrapper[24928]: I1205 11:05:21.417998 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7wp8\" (UniqueName: \"kubernetes.io/projected/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-kube-api-access-v7wp8\") pod \"cinder-72f7-account-create-update-rb6rq\" (UID: \"1c3b42bc-736b-4910-9bef-39ebeba0c5dd\") " pod="openstack/cinder-72f7-account-create-update-rb6rq" Dec 05 11:05:21.421628 master-0 kubenswrapper[24928]: I1205 11:05:21.420077 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kf7cj" Dec 05 11:05:21.421628 master-0 kubenswrapper[24928]: I1205 11:05:21.420241 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-operator-scripts\") pod \"cinder-72f7-account-create-update-rb6rq\" (UID: \"1c3b42bc-736b-4910-9bef-39ebeba0c5dd\") " pod="openstack/cinder-72f7-account-create-update-rb6rq" Dec 05 11:05:21.432267 master-0 kubenswrapper[24928]: I1205 11:05:21.432204 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-5m2fn"] Dec 05 11:05:21.433709 master-0 kubenswrapper[24928]: I1205 11:05:21.433677 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5m2fn" Dec 05 11:05:21.450148 master-0 kubenswrapper[24928]: I1205 11:05:21.447553 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7wp8\" (UniqueName: \"kubernetes.io/projected/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-kube-api-access-v7wp8\") pod \"cinder-72f7-account-create-update-rb6rq\" (UID: \"1c3b42bc-736b-4910-9bef-39ebeba0c5dd\") " pod="openstack/cinder-72f7-account-create-update-rb6rq" Dec 05 11:05:21.455800 master-0 kubenswrapper[24928]: I1205 11:05:21.455744 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-5m2fn"] Dec 05 11:05:21.459159 master-0 kubenswrapper[24928]: I1205 11:05:21.457608 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72f7-account-create-update-rb6rq" Dec 05 11:05:21.528360 master-0 kubenswrapper[24928]: I1205 11:05:21.521783 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1af478de-35a0-4480-a9cd-a2fa1a7ab774-operator-scripts\") pod \"neutron-db-create-5m2fn\" (UID: \"1af478de-35a0-4480-a9cd-a2fa1a7ab774\") " pod="openstack/neutron-db-create-5m2fn" Dec 05 11:05:21.528360 master-0 kubenswrapper[24928]: I1205 11:05:21.521941 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj49f\" (UniqueName: \"kubernetes.io/projected/1af478de-35a0-4480-a9cd-a2fa1a7ab774-kube-api-access-mj49f\") pod \"neutron-db-create-5m2fn\" (UID: \"1af478de-35a0-4480-a9cd-a2fa1a7ab774\") " pod="openstack/neutron-db-create-5m2fn" Dec 05 11:05:21.634452 master-0 kubenswrapper[24928]: I1205 11:05:21.631000 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj49f\" (UniqueName: \"kubernetes.io/projected/1af478de-35a0-4480-a9cd-a2fa1a7ab774-kube-api-access-mj49f\") pod \"neutron-db-create-5m2fn\" (UID: \"1af478de-35a0-4480-a9cd-a2fa1a7ab774\") " pod="openstack/neutron-db-create-5m2fn" Dec 05 11:05:21.634452 master-0 kubenswrapper[24928]: I1205 11:05:21.631242 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1af478de-35a0-4480-a9cd-a2fa1a7ab774-operator-scripts\") pod \"neutron-db-create-5m2fn\" (UID: \"1af478de-35a0-4480-a9cd-a2fa1a7ab774\") " pod="openstack/neutron-db-create-5m2fn" Dec 05 11:05:21.634452 master-0 kubenswrapper[24928]: I1205 11:05:21.632095 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1af478de-35a0-4480-a9cd-a2fa1a7ab774-operator-scripts\") pod \"neutron-db-create-5m2fn\" (UID: \"1af478de-35a0-4480-a9cd-a2fa1a7ab774\") " pod="openstack/neutron-db-create-5m2fn" Dec 05 11:05:21.846024 master-0 kubenswrapper[24928]: I1205 11:05:21.845965 24928 generic.go:334] "Generic (PLEG): container finished" podID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerID="40ffa0c10cc39cb58eb889575620e96ab5d9b9a2153e361fe6094e4d63c6ff83" exitCode=0 Dec 05 11:05:21.846341 master-0 kubenswrapper[24928]: I1205 11:05:21.846058 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdmr2" event={"ID":"98a4e639-cc2c-406e-b1b7-48746557cdc4","Type":"ContainerDied","Data":"40ffa0c10cc39cb58eb889575620e96ab5d9b9a2153e361fe6094e4d63c6ff83"} Dec 05 11:05:21.849193 master-0 kubenswrapper[24928]: I1205 11:05:21.849134 24928 generic.go:334] "Generic (PLEG): container finished" podID="032f0939-20e7-46cb-9665-b1346b7e1b10" containerID="199e2ba413ddeaeb33aee60c1f864ea1e644ea3137b0cc9fa2b3d04f4918b665" exitCode=0 Dec 05 11:05:21.850721 master-0 kubenswrapper[24928]: I1205 11:05:21.850623 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5nq4" event={"ID":"032f0939-20e7-46cb-9665-b1346b7e1b10","Type":"ContainerDied","Data":"199e2ba413ddeaeb33aee60c1f864ea1e644ea3137b0cc9fa2b3d04f4918b665"} Dec 05 11:05:22.148470 master-0 kubenswrapper[24928]: I1205 11:05:22.147629 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d29-account-create-update-dz2qv"] Dec 05 11:05:22.151537 master-0 kubenswrapper[24928]: I1205 11:05:22.149131 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d29-account-create-update-dz2qv" Dec 05 11:05:22.153841 master-0 kubenswrapper[24928]: I1205 11:05:22.153515 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 05 11:05:22.185311 master-0 kubenswrapper[24928]: I1205 11:05:22.183398 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d29-account-create-update-dz2qv"] Dec 05 11:05:22.193087 master-0 kubenswrapper[24928]: I1205 11:05:22.192368 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj49f\" (UniqueName: \"kubernetes.io/projected/1af478de-35a0-4480-a9cd-a2fa1a7ab774-kube-api-access-mj49f\") pod \"neutron-db-create-5m2fn\" (UID: \"1af478de-35a0-4480-a9cd-a2fa1a7ab774\") " pod="openstack/neutron-db-create-5m2fn" Dec 05 11:05:22.247680 master-0 kubenswrapper[24928]: I1205 11:05:22.247601 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxsmp\" (UniqueName: \"kubernetes.io/projected/0da83455-cb84-45a3-bbe7-211be81f4b56-kube-api-access-bxsmp\") pod \"neutron-5d29-account-create-update-dz2qv\" (UID: \"0da83455-cb84-45a3-bbe7-211be81f4b56\") " pod="openstack/neutron-5d29-account-create-update-dz2qv" Dec 05 11:05:22.247900 master-0 kubenswrapper[24928]: I1205 11:05:22.247690 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0da83455-cb84-45a3-bbe7-211be81f4b56-operator-scripts\") pod \"neutron-5d29-account-create-update-dz2qv\" (UID: \"0da83455-cb84-45a3-bbe7-211be81f4b56\") " pod="openstack/neutron-5d29-account-create-update-dz2qv" Dec 05 11:05:22.319247 master-0 kubenswrapper[24928]: I1205 11:05:22.319150 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fe5c439-67b4-4961-a197-736e934f5950" path="/var/lib/kubelet/pods/8fe5c439-67b4-4961-a197-736e934f5950/volumes" Dec 05 11:05:22.349746 master-0 kubenswrapper[24928]: I1205 11:05:22.345310 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-kf7cj"] Dec 05 11:05:22.356455 master-0 kubenswrapper[24928]: I1205 11:05:22.351096 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxsmp\" (UniqueName: \"kubernetes.io/projected/0da83455-cb84-45a3-bbe7-211be81f4b56-kube-api-access-bxsmp\") pod \"neutron-5d29-account-create-update-dz2qv\" (UID: \"0da83455-cb84-45a3-bbe7-211be81f4b56\") " pod="openstack/neutron-5d29-account-create-update-dz2qv" Dec 05 11:05:22.356455 master-0 kubenswrapper[24928]: I1205 11:05:22.351199 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0da83455-cb84-45a3-bbe7-211be81f4b56-operator-scripts\") pod \"neutron-5d29-account-create-update-dz2qv\" (UID: \"0da83455-cb84-45a3-bbe7-211be81f4b56\") " pod="openstack/neutron-5d29-account-create-update-dz2qv" Dec 05 11:05:22.356455 master-0 kubenswrapper[24928]: I1205 11:05:22.352123 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0da83455-cb84-45a3-bbe7-211be81f4b56-operator-scripts\") pod \"neutron-5d29-account-create-update-dz2qv\" (UID: \"0da83455-cb84-45a3-bbe7-211be81f4b56\") " pod="openstack/neutron-5d29-account-create-update-dz2qv" Dec 05 11:05:22.376936 master-0 kubenswrapper[24928]: I1205 11:05:22.376863 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-n74vh"] Dec 05 11:05:22.379355 master-0 kubenswrapper[24928]: I1205 11:05:22.379228 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxsmp\" (UniqueName: \"kubernetes.io/projected/0da83455-cb84-45a3-bbe7-211be81f4b56-kube-api-access-bxsmp\") pod \"neutron-5d29-account-create-update-dz2qv\" (UID: \"0da83455-cb84-45a3-bbe7-211be81f4b56\") " pod="openstack/neutron-5d29-account-create-update-dz2qv" Dec 05 11:05:22.379692 master-0 kubenswrapper[24928]: I1205 11:05:22.379652 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:22.382551 master-0 kubenswrapper[24928]: I1205 11:05:22.382511 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 11:05:22.382947 master-0 kubenswrapper[24928]: I1205 11:05:22.382926 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 11:05:22.383078 master-0 kubenswrapper[24928]: I1205 11:05:22.383058 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 11:05:22.390244 master-0 kubenswrapper[24928]: I1205 11:05:22.389769 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5m2fn" Dec 05 11:05:22.410625 master-0 kubenswrapper[24928]: I1205 11:05:22.410556 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-n74vh"] Dec 05 11:05:22.422186 master-0 kubenswrapper[24928]: I1205 11:05:22.422137 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-72f7-account-create-update-rb6rq"] Dec 05 11:05:22.453521 master-0 kubenswrapper[24928]: I1205 11:05:22.453458 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krpgx\" (UniqueName: \"kubernetes.io/projected/f038c921-52be-46ef-ad30-190252e42dcb-kube-api-access-krpgx\") pod \"keystone-db-sync-n74vh\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:22.453750 master-0 kubenswrapper[24928]: I1205 11:05:22.453669 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-combined-ca-bundle\") pod \"keystone-db-sync-n74vh\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:22.453750 master-0 kubenswrapper[24928]: I1205 11:05:22.453717 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-config-data\") pod \"keystone-db-sync-n74vh\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:22.507565 master-0 kubenswrapper[24928]: I1205 11:05:22.506335 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d29-account-create-update-dz2qv" Dec 05 11:05:22.555348 master-0 kubenswrapper[24928]: I1205 11:05:22.555273 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-combined-ca-bundle\") pod \"keystone-db-sync-n74vh\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:22.555622 master-0 kubenswrapper[24928]: I1205 11:05:22.555389 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-config-data\") pod \"keystone-db-sync-n74vh\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:22.556901 master-0 kubenswrapper[24928]: I1205 11:05:22.556872 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krpgx\" (UniqueName: \"kubernetes.io/projected/f038c921-52be-46ef-ad30-190252e42dcb-kube-api-access-krpgx\") pod \"keystone-db-sync-n74vh\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:22.559324 master-0 kubenswrapper[24928]: I1205 11:05:22.559283 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-combined-ca-bundle\") pod \"keystone-db-sync-n74vh\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:22.564597 master-0 kubenswrapper[24928]: I1205 11:05:22.564538 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-config-data\") pod \"keystone-db-sync-n74vh\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:22.583757 master-0 kubenswrapper[24928]: I1205 11:05:22.582914 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krpgx\" (UniqueName: \"kubernetes.io/projected/f038c921-52be-46ef-ad30-190252e42dcb-kube-api-access-krpgx\") pod \"keystone-db-sync-n74vh\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:22.591271 master-0 kubenswrapper[24928]: I1205 11:05:22.590589 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:22.877755 master-0 kubenswrapper[24928]: I1205 11:05:22.877672 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-72f7-account-create-update-rb6rq" event={"ID":"1c3b42bc-736b-4910-9bef-39ebeba0c5dd","Type":"ContainerStarted","Data":"fd71a617a2a9537d666f3f1e92830e03cf7fdbfacf21b753864f4557ebfcb901"} Dec 05 11:05:22.880023 master-0 kubenswrapper[24928]: I1205 11:05:22.879974 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kf7cj" event={"ID":"17f4e63e-29f0-4884-a20e-a3cb02c85235","Type":"ContainerStarted","Data":"8fd024143bfad9fc15feed56164eef4751f867f58dbfdcb5bd9d168492a6679f"} Dec 05 11:05:22.882544 master-0 kubenswrapper[24928]: I1205 11:05:22.882462 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"ad0afe60ebfc6442dc48585f53794d904a0e78fce3a29844684de3c3f603534d"} Dec 05 11:05:22.969831 master-0 kubenswrapper[24928]: I1205 11:05:22.969766 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-5m2fn"] Dec 05 11:05:22.980617 master-0 kubenswrapper[24928]: W1205 11:05:22.980262 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1af478de_35a0_4480_a9cd_a2fa1a7ab774.slice/crio-c5f7f3dfd6ceb2e46d48246ec9b17447de1c8bbb05f25f6795ffaef3231372db WatchSource:0}: Error finding container c5f7f3dfd6ceb2e46d48246ec9b17447de1c8bbb05f25f6795ffaef3231372db: Status 404 returned error can't find the container with id c5f7f3dfd6ceb2e46d48246ec9b17447de1c8bbb05f25f6795ffaef3231372db Dec 05 11:05:23.015628 master-0 kubenswrapper[24928]: I1205 11:05:23.015553 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lvktj"] Dec 05 11:05:23.016242 master-0 kubenswrapper[24928]: I1205 11:05:23.016026 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lvktj" podUID="0e07646a-fd0f-49fc-89da-5938c26aad18" containerName="registry-server" containerID="cri-o://9cf3b503a6c68397953696990b7e6646d8f3a7dd04827ef76e79e29506ae342a" gracePeriod=2 Dec 05 11:05:23.130825 master-0 kubenswrapper[24928]: I1205 11:05:23.130758 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d29-account-create-update-dz2qv"] Dec 05 11:05:23.198889 master-0 kubenswrapper[24928]: W1205 11:05:23.198842 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0da83455_cb84_45a3_bbe7_211be81f4b56.slice/crio-6a56082f38bcbb1f391fa87ae1c99041a6892e70af8102232dee1023386ddf4b WatchSource:0}: Error finding container 6a56082f38bcbb1f391fa87ae1c99041a6892e70af8102232dee1023386ddf4b: Status 404 returned error can't find the container with id 6a56082f38bcbb1f391fa87ae1c99041a6892e70af8102232dee1023386ddf4b Dec 05 11:05:23.680837 master-0 kubenswrapper[24928]: I1205 11:05:23.680735 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-n74vh"] Dec 05 11:05:23.899205 master-0 kubenswrapper[24928]: I1205 11:05:23.899124 24928 generic.go:334] "Generic (PLEG): container finished" podID="0e07646a-fd0f-49fc-89da-5938c26aad18" containerID="9cf3b503a6c68397953696990b7e6646d8f3a7dd04827ef76e79e29506ae342a" exitCode=0 Dec 05 11:05:23.899468 master-0 kubenswrapper[24928]: I1205 11:05:23.899228 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvktj" event={"ID":"0e07646a-fd0f-49fc-89da-5938c26aad18","Type":"ContainerDied","Data":"9cf3b503a6c68397953696990b7e6646d8f3a7dd04827ef76e79e29506ae342a"} Dec 05 11:05:23.902538 master-0 kubenswrapper[24928]: I1205 11:05:23.902489 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdmr2" event={"ID":"98a4e639-cc2c-406e-b1b7-48746557cdc4","Type":"ContainerStarted","Data":"465eaaacdeaddc9af9ecd8ff210c679f3a485340cb493223b88c974544b2b75d"} Dec 05 11:05:23.908035 master-0 kubenswrapper[24928]: I1205 11:05:23.907971 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5m2fn" event={"ID":"1af478de-35a0-4480-a9cd-a2fa1a7ab774","Type":"ContainerStarted","Data":"0c8d87e172238ddbcec7bd55d985c51a25f017fae7c7d331c80fbbbad8c2b0f7"} Dec 05 11:05:23.908269 master-0 kubenswrapper[24928]: I1205 11:05:23.908248 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5m2fn" event={"ID":"1af478de-35a0-4480-a9cd-a2fa1a7ab774","Type":"ContainerStarted","Data":"c5f7f3dfd6ceb2e46d48246ec9b17447de1c8bbb05f25f6795ffaef3231372db"} Dec 05 11:05:23.913622 master-0 kubenswrapper[24928]: I1205 11:05:23.913564 24928 generic.go:334] "Generic (PLEG): container finished" podID="1c3b42bc-736b-4910-9bef-39ebeba0c5dd" containerID="bdcd9e7bb6583c8e0cbcf590d5ffdb517083382508c79285da8340efa4efe568" exitCode=0 Dec 05 11:05:23.913952 master-0 kubenswrapper[24928]: I1205 11:05:23.913923 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-72f7-account-create-update-rb6rq" event={"ID":"1c3b42bc-736b-4910-9bef-39ebeba0c5dd","Type":"ContainerDied","Data":"bdcd9e7bb6583c8e0cbcf590d5ffdb517083382508c79285da8340efa4efe568"} Dec 05 11:05:23.917983 master-0 kubenswrapper[24928]: I1205 11:05:23.917917 24928 generic.go:334] "Generic (PLEG): container finished" podID="17f4e63e-29f0-4884-a20e-a3cb02c85235" containerID="6c94531a9e45f78a5382d174371e1bcb6e1cbbfbbca55c1ca06780e0f1c6d491" exitCode=0 Dec 05 11:05:23.918130 master-0 kubenswrapper[24928]: I1205 11:05:23.918071 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kf7cj" event={"ID":"17f4e63e-29f0-4884-a20e-a3cb02c85235","Type":"ContainerDied","Data":"6c94531a9e45f78a5382d174371e1bcb6e1cbbfbbca55c1ca06780e0f1c6d491"} Dec 05 11:05:23.921585 master-0 kubenswrapper[24928]: I1205 11:05:23.921306 24928 generic.go:334] "Generic (PLEG): container finished" podID="032f0939-20e7-46cb-9665-b1346b7e1b10" containerID="35606e6245a10ef961fcc399697eab348e9c687982934c3772679dc5994ccbb8" exitCode=0 Dec 05 11:05:23.921585 master-0 kubenswrapper[24928]: I1205 11:05:23.921404 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5nq4" event={"ID":"032f0939-20e7-46cb-9665-b1346b7e1b10","Type":"ContainerDied","Data":"35606e6245a10ef961fcc399697eab348e9c687982934c3772679dc5994ccbb8"} Dec 05 11:05:23.928962 master-0 kubenswrapper[24928]: I1205 11:05:23.927768 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d29-account-create-update-dz2qv" event={"ID":"0da83455-cb84-45a3-bbe7-211be81f4b56","Type":"ContainerStarted","Data":"8b6b32e3398efe0eb9fdc0d0519779f30f19812a641a6dead4c0ffda77141b77"} Dec 05 11:05:23.928962 master-0 kubenswrapper[24928]: I1205 11:05:23.927860 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d29-account-create-update-dz2qv" event={"ID":"0da83455-cb84-45a3-bbe7-211be81f4b56","Type":"ContainerStarted","Data":"6a56082f38bcbb1f391fa87ae1c99041a6892e70af8102232dee1023386ddf4b"} Dec 05 11:05:24.013157 master-0 kubenswrapper[24928]: I1205 11:05:24.011992 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d29-account-create-update-dz2qv" podStartSLOduration=3.011971286 podStartE2EDuration="3.011971286s" podCreationTimestamp="2025-12-05 11:05:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:05:23.995318661 +0000 UTC m=+1083.998512522" watchObservedRunningTime="2025-12-05 11:05:24.011971286 +0000 UTC m=+1084.015165137" Dec 05 11:05:24.041357 master-0 kubenswrapper[24928]: I1205 11:05:24.041264 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-5m2fn" podStartSLOduration=3.041240826 podStartE2EDuration="3.041240826s" podCreationTimestamp="2025-12-05 11:05:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:05:24.034353064 +0000 UTC m=+1084.037546915" watchObservedRunningTime="2025-12-05 11:05:24.041240826 +0000 UTC m=+1084.044434697" Dec 05 11:05:24.110066 master-0 kubenswrapper[24928]: W1205 11:05:24.109373 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf038c921_52be_46ef_ad30_190252e42dcb.slice/crio-89910f1088d283452982e26d014a8925a51471c84e35192ca31ec95992bfefe8 WatchSource:0}: Error finding container 89910f1088d283452982e26d014a8925a51471c84e35192ca31ec95992bfefe8: Status 404 returned error can't find the container with id 89910f1088d283452982e26d014a8925a51471c84e35192ca31ec95992bfefe8 Dec 05 11:05:24.169293 master-0 kubenswrapper[24928]: I1205 11:05:24.169237 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:24.254963 master-0 kubenswrapper[24928]: I1205 11:05:24.254893 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-catalog-content\") pod \"0e07646a-fd0f-49fc-89da-5938c26aad18\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " Dec 05 11:05:24.255746 master-0 kubenswrapper[24928]: I1205 11:05:24.255205 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqwh7\" (UniqueName: \"kubernetes.io/projected/0e07646a-fd0f-49fc-89da-5938c26aad18-kube-api-access-pqwh7\") pod \"0e07646a-fd0f-49fc-89da-5938c26aad18\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " Dec 05 11:05:24.255746 master-0 kubenswrapper[24928]: I1205 11:05:24.255386 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-utilities\") pod \"0e07646a-fd0f-49fc-89da-5938c26aad18\" (UID: \"0e07646a-fd0f-49fc-89da-5938c26aad18\") " Dec 05 11:05:24.256819 master-0 kubenswrapper[24928]: I1205 11:05:24.256760 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-utilities" (OuterVolumeSpecName: "utilities") pod "0e07646a-fd0f-49fc-89da-5938c26aad18" (UID: "0e07646a-fd0f-49fc-89da-5938c26aad18"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:05:24.258728 master-0 kubenswrapper[24928]: I1205 11:05:24.258685 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e07646a-fd0f-49fc-89da-5938c26aad18-kube-api-access-pqwh7" (OuterVolumeSpecName: "kube-api-access-pqwh7") pod "0e07646a-fd0f-49fc-89da-5938c26aad18" (UID: "0e07646a-fd0f-49fc-89da-5938c26aad18"). InnerVolumeSpecName "kube-api-access-pqwh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:24.272063 master-0 kubenswrapper[24928]: I1205 11:05:24.270878 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e07646a-fd0f-49fc-89da-5938c26aad18" (UID: "0e07646a-fd0f-49fc-89da-5938c26aad18"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:05:24.357840 master-0 kubenswrapper[24928]: I1205 11:05:24.357760 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqwh7\" (UniqueName: \"kubernetes.io/projected/0e07646a-fd0f-49fc-89da-5938c26aad18-kube-api-access-pqwh7\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:24.357840 master-0 kubenswrapper[24928]: I1205 11:05:24.357814 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:24.357840 master-0 kubenswrapper[24928]: I1205 11:05:24.357829 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e07646a-fd0f-49fc-89da-5938c26aad18-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:24.953745 master-0 kubenswrapper[24928]: I1205 11:05:24.953616 24928 generic.go:334] "Generic (PLEG): container finished" podID="0da83455-cb84-45a3-bbe7-211be81f4b56" containerID="8b6b32e3398efe0eb9fdc0d0519779f30f19812a641a6dead4c0ffda77141b77" exitCode=0 Dec 05 11:05:24.953745 master-0 kubenswrapper[24928]: I1205 11:05:24.953746 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d29-account-create-update-dz2qv" event={"ID":"0da83455-cb84-45a3-bbe7-211be81f4b56","Type":"ContainerDied","Data":"8b6b32e3398efe0eb9fdc0d0519779f30f19812a641a6dead4c0ffda77141b77"} Dec 05 11:05:24.957053 master-0 kubenswrapper[24928]: I1205 11:05:24.957007 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-n74vh" event={"ID":"f038c921-52be-46ef-ad30-190252e42dcb","Type":"ContainerStarted","Data":"89910f1088d283452982e26d014a8925a51471c84e35192ca31ec95992bfefe8"} Dec 05 11:05:24.960601 master-0 kubenswrapper[24928]: I1205 11:05:24.960474 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lvktj" event={"ID":"0e07646a-fd0f-49fc-89da-5938c26aad18","Type":"ContainerDied","Data":"97ed2f6a5693ad410a9131bea71f95a34435534fae8d925fcb7524ccba134f62"} Dec 05 11:05:24.960694 master-0 kubenswrapper[24928]: I1205 11:05:24.960625 24928 scope.go:117] "RemoveContainer" containerID="9cf3b503a6c68397953696990b7e6646d8f3a7dd04827ef76e79e29506ae342a" Dec 05 11:05:24.960694 master-0 kubenswrapper[24928]: I1205 11:05:24.960492 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lvktj" Dec 05 11:05:24.965123 master-0 kubenswrapper[24928]: I1205 11:05:24.965042 24928 generic.go:334] "Generic (PLEG): container finished" podID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerID="465eaaacdeaddc9af9ecd8ff210c679f3a485340cb493223b88c974544b2b75d" exitCode=0 Dec 05 11:05:24.965242 master-0 kubenswrapper[24928]: I1205 11:05:24.965164 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdmr2" event={"ID":"98a4e639-cc2c-406e-b1b7-48746557cdc4","Type":"ContainerDied","Data":"465eaaacdeaddc9af9ecd8ff210c679f3a485340cb493223b88c974544b2b75d"} Dec 05 11:05:24.969473 master-0 kubenswrapper[24928]: I1205 11:05:24.969368 24928 generic.go:334] "Generic (PLEG): container finished" podID="1af478de-35a0-4480-a9cd-a2fa1a7ab774" containerID="0c8d87e172238ddbcec7bd55d985c51a25f017fae7c7d331c80fbbbad8c2b0f7" exitCode=0 Dec 05 11:05:24.969575 master-0 kubenswrapper[24928]: I1205 11:05:24.969495 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5m2fn" event={"ID":"1af478de-35a0-4480-a9cd-a2fa1a7ab774","Type":"ContainerDied","Data":"0c8d87e172238ddbcec7bd55d985c51a25f017fae7c7d331c80fbbbad8c2b0f7"} Dec 05 11:05:27.924823 master-0 kubenswrapper[24928]: I1205 11:05:27.922790 24928 scope.go:117] "RemoveContainer" containerID="d66ef34a556dbec95523b1b94b6f318111a67e1318d2494a1fff969cd14b67dc" Dec 05 11:05:28.071279 master-0 kubenswrapper[24928]: I1205 11:05:28.071212 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kf7cj" Dec 05 11:05:28.071566 master-0 kubenswrapper[24928]: I1205 11:05:28.071482 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-5m2fn" event={"ID":"1af478de-35a0-4480-a9cd-a2fa1a7ab774","Type":"ContainerDied","Data":"c5f7f3dfd6ceb2e46d48246ec9b17447de1c8bbb05f25f6795ffaef3231372db"} Dec 05 11:05:28.071566 master-0 kubenswrapper[24928]: I1205 11:05:28.071518 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5f7f3dfd6ceb2e46d48246ec9b17447de1c8bbb05f25f6795ffaef3231372db" Dec 05 11:05:28.113370 master-0 kubenswrapper[24928]: I1205 11:05:28.112643 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lvktj"] Dec 05 11:05:28.113848 master-0 kubenswrapper[24928]: I1205 11:05:28.113797 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-72f7-account-create-update-rb6rq" event={"ID":"1c3b42bc-736b-4910-9bef-39ebeba0c5dd","Type":"ContainerDied","Data":"fd71a617a2a9537d666f3f1e92830e03cf7fdbfacf21b753864f4557ebfcb901"} Dec 05 11:05:28.113902 master-0 kubenswrapper[24928]: I1205 11:05:28.113853 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd71a617a2a9537d666f3f1e92830e03cf7fdbfacf21b753864f4557ebfcb901" Dec 05 11:05:28.120932 master-0 kubenswrapper[24928]: I1205 11:05:28.120856 24928 scope.go:117] "RemoveContainer" containerID="e7546ce87acba6b8e0104bd03704f38f94b6ae6c375b852b6cc252028b5419ab" Dec 05 11:05:28.131477 master-0 kubenswrapper[24928]: I1205 11:05:28.125383 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-kf7cj" event={"ID":"17f4e63e-29f0-4884-a20e-a3cb02c85235","Type":"ContainerDied","Data":"8fd024143bfad9fc15feed56164eef4751f867f58dbfdcb5bd9d168492a6679f"} Dec 05 11:05:28.131477 master-0 kubenswrapper[24928]: I1205 11:05:28.125469 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fd024143bfad9fc15feed56164eef4751f867f58dbfdcb5bd9d168492a6679f" Dec 05 11:05:28.131477 master-0 kubenswrapper[24928]: I1205 11:05:28.125546 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-kf7cj" Dec 05 11:05:28.132144 master-0 kubenswrapper[24928]: I1205 11:05:28.131950 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d29-account-create-update-dz2qv" event={"ID":"0da83455-cb84-45a3-bbe7-211be81f4b56","Type":"ContainerDied","Data":"6a56082f38bcbb1f391fa87ae1c99041a6892e70af8102232dee1023386ddf4b"} Dec 05 11:05:28.132144 master-0 kubenswrapper[24928]: I1205 11:05:28.132009 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a56082f38bcbb1f391fa87ae1c99041a6892e70af8102232dee1023386ddf4b" Dec 05 11:05:28.158384 master-0 kubenswrapper[24928]: I1205 11:05:28.156646 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17f4e63e-29f0-4884-a20e-a3cb02c85235-operator-scripts\") pod \"17f4e63e-29f0-4884-a20e-a3cb02c85235\" (UID: \"17f4e63e-29f0-4884-a20e-a3cb02c85235\") " Dec 05 11:05:28.158384 master-0 kubenswrapper[24928]: I1205 11:05:28.156998 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wtqqt\" (UniqueName: \"kubernetes.io/projected/17f4e63e-29f0-4884-a20e-a3cb02c85235-kube-api-access-wtqqt\") pod \"17f4e63e-29f0-4884-a20e-a3cb02c85235\" (UID: \"17f4e63e-29f0-4884-a20e-a3cb02c85235\") " Dec 05 11:05:28.158384 master-0 kubenswrapper[24928]: I1205 11:05:28.158005 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17f4e63e-29f0-4884-a20e-a3cb02c85235-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "17f4e63e-29f0-4884-a20e-a3cb02c85235" (UID: "17f4e63e-29f0-4884-a20e-a3cb02c85235"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:28.164112 master-0 kubenswrapper[24928]: I1205 11:05:28.163762 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17f4e63e-29f0-4884-a20e-a3cb02c85235-kube-api-access-wtqqt" (OuterVolumeSpecName: "kube-api-access-wtqqt") pod "17f4e63e-29f0-4884-a20e-a3cb02c85235" (UID: "17f4e63e-29f0-4884-a20e-a3cb02c85235"). InnerVolumeSpecName "kube-api-access-wtqqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:28.176183 master-0 kubenswrapper[24928]: I1205 11:05:28.176119 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lvktj"] Dec 05 11:05:28.194670 master-0 kubenswrapper[24928]: I1205 11:05:28.194595 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72f7-account-create-update-rb6rq" Dec 05 11:05:28.259016 master-0 kubenswrapper[24928]: I1205 11:05:28.256781 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e07646a-fd0f-49fc-89da-5938c26aad18" path="/var/lib/kubelet/pods/0e07646a-fd0f-49fc-89da-5938c26aad18/volumes" Dec 05 11:05:28.260949 master-0 kubenswrapper[24928]: I1205 11:05:28.260903 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wtqqt\" (UniqueName: \"kubernetes.io/projected/17f4e63e-29f0-4884-a20e-a3cb02c85235-kube-api-access-wtqqt\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:28.260949 master-0 kubenswrapper[24928]: I1205 11:05:28.260947 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/17f4e63e-29f0-4884-a20e-a3cb02c85235-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:28.271699 master-0 kubenswrapper[24928]: I1205 11:05:28.270404 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5m2fn" Dec 05 11:05:28.327713 master-0 kubenswrapper[24928]: I1205 11:05:28.327604 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d29-account-create-update-dz2qv" Dec 05 11:05:28.374227 master-0 kubenswrapper[24928]: I1205 11:05:28.364628 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj49f\" (UniqueName: \"kubernetes.io/projected/1af478de-35a0-4480-a9cd-a2fa1a7ab774-kube-api-access-mj49f\") pod \"1af478de-35a0-4480-a9cd-a2fa1a7ab774\" (UID: \"1af478de-35a0-4480-a9cd-a2fa1a7ab774\") " Dec 05 11:05:28.374227 master-0 kubenswrapper[24928]: I1205 11:05:28.364845 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7wp8\" (UniqueName: \"kubernetes.io/projected/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-kube-api-access-v7wp8\") pod \"1c3b42bc-736b-4910-9bef-39ebeba0c5dd\" (UID: \"1c3b42bc-736b-4910-9bef-39ebeba0c5dd\") " Dec 05 11:05:28.374227 master-0 kubenswrapper[24928]: I1205 11:05:28.365023 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-operator-scripts\") pod \"1c3b42bc-736b-4910-9bef-39ebeba0c5dd\" (UID: \"1c3b42bc-736b-4910-9bef-39ebeba0c5dd\") " Dec 05 11:05:28.374227 master-0 kubenswrapper[24928]: I1205 11:05:28.365100 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1af478de-35a0-4480-a9cd-a2fa1a7ab774-operator-scripts\") pod \"1af478de-35a0-4480-a9cd-a2fa1a7ab774\" (UID: \"1af478de-35a0-4480-a9cd-a2fa1a7ab774\") " Dec 05 11:05:28.374227 master-0 kubenswrapper[24928]: I1205 11:05:28.369734 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1af478de-35a0-4480-a9cd-a2fa1a7ab774-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1af478de-35a0-4480-a9cd-a2fa1a7ab774" (UID: "1af478de-35a0-4480-a9cd-a2fa1a7ab774"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:28.374227 master-0 kubenswrapper[24928]: I1205 11:05:28.370579 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1af478de-35a0-4480-a9cd-a2fa1a7ab774-kube-api-access-mj49f" (OuterVolumeSpecName: "kube-api-access-mj49f") pod "1af478de-35a0-4480-a9cd-a2fa1a7ab774" (UID: "1af478de-35a0-4480-a9cd-a2fa1a7ab774"). InnerVolumeSpecName "kube-api-access-mj49f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:28.374227 master-0 kubenswrapper[24928]: I1205 11:05:28.371019 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1c3b42bc-736b-4910-9bef-39ebeba0c5dd" (UID: "1c3b42bc-736b-4910-9bef-39ebeba0c5dd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:28.374227 master-0 kubenswrapper[24928]: I1205 11:05:28.372205 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-kube-api-access-v7wp8" (OuterVolumeSpecName: "kube-api-access-v7wp8") pod "1c3b42bc-736b-4910-9bef-39ebeba0c5dd" (UID: "1c3b42bc-736b-4910-9bef-39ebeba0c5dd"). InnerVolumeSpecName "kube-api-access-v7wp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:28.469467 master-0 kubenswrapper[24928]: I1205 11:05:28.468543 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0da83455-cb84-45a3-bbe7-211be81f4b56-operator-scripts\") pod \"0da83455-cb84-45a3-bbe7-211be81f4b56\" (UID: \"0da83455-cb84-45a3-bbe7-211be81f4b56\") " Dec 05 11:05:28.469467 master-0 kubenswrapper[24928]: I1205 11:05:28.469120 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxsmp\" (UniqueName: \"kubernetes.io/projected/0da83455-cb84-45a3-bbe7-211be81f4b56-kube-api-access-bxsmp\") pod \"0da83455-cb84-45a3-bbe7-211be81f4b56\" (UID: \"0da83455-cb84-45a3-bbe7-211be81f4b56\") " Dec 05 11:05:28.469621 master-0 kubenswrapper[24928]: I1205 11:05:28.468993 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0da83455-cb84-45a3-bbe7-211be81f4b56-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0da83455-cb84-45a3-bbe7-211be81f4b56" (UID: "0da83455-cb84-45a3-bbe7-211be81f4b56"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:28.469621 master-0 kubenswrapper[24928]: I1205 11:05:28.469610 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj49f\" (UniqueName: \"kubernetes.io/projected/1af478de-35a0-4480-a9cd-a2fa1a7ab774-kube-api-access-mj49f\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:28.469686 master-0 kubenswrapper[24928]: I1205 11:05:28.469628 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7wp8\" (UniqueName: \"kubernetes.io/projected/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-kube-api-access-v7wp8\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:28.469686 master-0 kubenswrapper[24928]: I1205 11:05:28.469640 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1c3b42bc-736b-4910-9bef-39ebeba0c5dd-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:28.469686 master-0 kubenswrapper[24928]: I1205 11:05:28.469649 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1af478de-35a0-4480-a9cd-a2fa1a7ab774-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:28.475319 master-0 kubenswrapper[24928]: I1205 11:05:28.475229 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0da83455-cb84-45a3-bbe7-211be81f4b56-kube-api-access-bxsmp" (OuterVolumeSpecName: "kube-api-access-bxsmp") pod "0da83455-cb84-45a3-bbe7-211be81f4b56" (UID: "0da83455-cb84-45a3-bbe7-211be81f4b56"). InnerVolumeSpecName "kube-api-access-bxsmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:28.578104 master-0 kubenswrapper[24928]: I1205 11:05:28.577991 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0da83455-cb84-45a3-bbe7-211be81f4b56-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:28.578104 master-0 kubenswrapper[24928]: I1205 11:05:28.578057 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxsmp\" (UniqueName: \"kubernetes.io/projected/0da83455-cb84-45a3-bbe7-211be81f4b56-kube-api-access-bxsmp\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:29.149962 master-0 kubenswrapper[24928]: I1205 11:05:29.149860 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5nq4" event={"ID":"032f0939-20e7-46cb-9665-b1346b7e1b10","Type":"ContainerStarted","Data":"1ad8d4221e10a10b242cfb51bce3c73c08c96b11c2a121dce15f2c8ce277ce37"} Dec 05 11:05:29.157120 master-0 kubenswrapper[24928]: I1205 11:05:29.157058 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"933d27078aca48dd660e6371ba198dd4ca01b8b4af98ed79a0e9841f9cfd76c3"} Dec 05 11:05:29.157120 master-0 kubenswrapper[24928]: I1205 11:05:29.157114 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"e30ab79911d4c873f8f3dd6c6dee5de290c7e37552f855af588b80e03b694746"} Dec 05 11:05:29.157120 master-0 kubenswrapper[24928]: I1205 11:05:29.157131 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"e2c3dc4591517cebf72b3b02cf206fc7394e545abc5aeb14aebf49d1b00d8257"} Dec 05 11:05:29.165641 master-0 kubenswrapper[24928]: I1205 11:05:29.165545 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdmr2" event={"ID":"98a4e639-cc2c-406e-b1b7-48746557cdc4","Type":"ContainerStarted","Data":"32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e"} Dec 05 11:05:29.165641 master-0 kubenswrapper[24928]: I1205 11:05:29.165576 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-5m2fn" Dec 05 11:05:29.166046 master-0 kubenswrapper[24928]: I1205 11:05:29.165724 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-72f7-account-create-update-rb6rq" Dec 05 11:05:29.167897 master-0 kubenswrapper[24928]: I1205 11:05:29.167841 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d29-account-create-update-dz2qv" Dec 05 11:05:29.922446 master-0 kubenswrapper[24928]: I1205 11:05:29.920908 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:29.922446 master-0 kubenswrapper[24928]: I1205 11:05:29.920993 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:29.935250 master-0 kubenswrapper[24928]: I1205 11:05:29.935148 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:29.935250 master-0 kubenswrapper[24928]: I1205 11:05:29.935233 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:29.955543 master-0 kubenswrapper[24928]: I1205 11:05:29.955356 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x5nq4" podStartSLOduration=5.685962676 podStartE2EDuration="11.95532295s" podCreationTimestamp="2025-12-05 11:05:18 +0000 UTC" firstStartedPulling="2025-12-05 11:05:21.851331858 +0000 UTC m=+1081.854525709" lastFinishedPulling="2025-12-05 11:05:28.120692132 +0000 UTC m=+1088.123885983" observedRunningTime="2025-12-05 11:05:29.861332986 +0000 UTC m=+1089.864526837" watchObservedRunningTime="2025-12-05 11:05:29.95532295 +0000 UTC m=+1089.958516791" Dec 05 11:05:29.982443 master-0 kubenswrapper[24928]: I1205 11:05:29.982260 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vdmr2" podStartSLOduration=5.472714786 podStartE2EDuration="11.98222459s" podCreationTimestamp="2025-12-05 11:05:18 +0000 UTC" firstStartedPulling="2025-12-05 11:05:21.847568054 +0000 UTC m=+1081.850761895" lastFinishedPulling="2025-12-05 11:05:28.357077858 +0000 UTC m=+1088.360271699" observedRunningTime="2025-12-05 11:05:29.91605559 +0000 UTC m=+1089.919249441" watchObservedRunningTime="2025-12-05 11:05:29.98222459 +0000 UTC m=+1089.985418471" Dec 05 11:05:30.004577 master-0 kubenswrapper[24928]: I1205 11:05:30.004455 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:30.181627 master-0 kubenswrapper[24928]: I1205 11:05:30.181536 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"c2b77e2b8cfcea361addd6702b2280c010bde3ed10420097836ef673d06f34d3"} Dec 05 11:05:31.004461 master-0 kubenswrapper[24928]: I1205 11:05:31.004369 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vdmr2" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerName="registry-server" probeResult="failure" output=< Dec 05 11:05:31.004461 master-0 kubenswrapper[24928]: timeout: failed to connect service ":50051" within 1s Dec 05 11:05:31.004461 master-0 kubenswrapper[24928]: > Dec 05 11:05:31.630340 master-0 kubenswrapper[24928]: I1205 11:05:31.194678 24928 generic.go:334] "Generic (PLEG): container finished" podID="1d10625d-a13a-4538-9d41-6e71f1af3669" containerID="3370f573d17f498b4ea25781e741647ab3a6193bb5b0f8a6e64abaa22c3dab7d" exitCode=0 Dec 05 11:05:31.630340 master-0 kubenswrapper[24928]: I1205 11:05:31.194784 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bp44h" event={"ID":"1d10625d-a13a-4538-9d41-6e71f1af3669","Type":"ContainerDied","Data":"3370f573d17f498b4ea25781e741647ab3a6193bb5b0f8a6e64abaa22c3dab7d"} Dec 05 11:05:32.924932 master-0 kubenswrapper[24928]: I1205 11:05:32.924124 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:33.027465 master-0 kubenswrapper[24928]: I1205 11:05:33.015202 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-combined-ca-bundle\") pod \"1d10625d-a13a-4538-9d41-6e71f1af3669\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " Dec 05 11:05:33.027465 master-0 kubenswrapper[24928]: I1205 11:05:33.015380 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-config-data\") pod \"1d10625d-a13a-4538-9d41-6e71f1af3669\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " Dec 05 11:05:33.027465 master-0 kubenswrapper[24928]: I1205 11:05:33.015645 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdqcv\" (UniqueName: \"kubernetes.io/projected/1d10625d-a13a-4538-9d41-6e71f1af3669-kube-api-access-gdqcv\") pod \"1d10625d-a13a-4538-9d41-6e71f1af3669\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " Dec 05 11:05:33.027465 master-0 kubenswrapper[24928]: I1205 11:05:33.015718 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-db-sync-config-data\") pod \"1d10625d-a13a-4538-9d41-6e71f1af3669\" (UID: \"1d10625d-a13a-4538-9d41-6e71f1af3669\") " Dec 05 11:05:33.027465 master-0 kubenswrapper[24928]: I1205 11:05:33.021198 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1d10625d-a13a-4538-9d41-6e71f1af3669" (UID: "1d10625d-a13a-4538-9d41-6e71f1af3669"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:33.027465 master-0 kubenswrapper[24928]: I1205 11:05:33.025633 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d10625d-a13a-4538-9d41-6e71f1af3669-kube-api-access-gdqcv" (OuterVolumeSpecName: "kube-api-access-gdqcv") pod "1d10625d-a13a-4538-9d41-6e71f1af3669" (UID: "1d10625d-a13a-4538-9d41-6e71f1af3669"). InnerVolumeSpecName "kube-api-access-gdqcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:33.058500 master-0 kubenswrapper[24928]: I1205 11:05:33.056781 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d10625d-a13a-4538-9d41-6e71f1af3669" (UID: "1d10625d-a13a-4538-9d41-6e71f1af3669"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:33.095852 master-0 kubenswrapper[24928]: I1205 11:05:33.095712 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-config-data" (OuterVolumeSpecName: "config-data") pod "1d10625d-a13a-4538-9d41-6e71f1af3669" (UID: "1d10625d-a13a-4538-9d41-6e71f1af3669"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:33.119697 master-0 kubenswrapper[24928]: I1205 11:05:33.119628 24928 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-db-sync-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:33.119697 master-0 kubenswrapper[24928]: I1205 11:05:33.119687 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:33.119697 master-0 kubenswrapper[24928]: I1205 11:05:33.119705 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d10625d-a13a-4538-9d41-6e71f1af3669-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:33.119998 master-0 kubenswrapper[24928]: I1205 11:05:33.119723 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdqcv\" (UniqueName: \"kubernetes.io/projected/1d10625d-a13a-4538-9d41-6e71f1af3669-kube-api-access-gdqcv\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:33.218475 master-0 kubenswrapper[24928]: I1205 11:05:33.218383 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-bp44h" event={"ID":"1d10625d-a13a-4538-9d41-6e71f1af3669","Type":"ContainerDied","Data":"c83fefe62f19dca5c51029a78a321b7f583eb37db5813c2105addbd16ee31f81"} Dec 05 11:05:33.218475 master-0 kubenswrapper[24928]: I1205 11:05:33.218459 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c83fefe62f19dca5c51029a78a321b7f583eb37db5813c2105addbd16ee31f81" Dec 05 11:05:33.218783 master-0 kubenswrapper[24928]: I1205 11:05:33.218523 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-bp44h" Dec 05 11:05:33.230893 master-0 kubenswrapper[24928]: I1205 11:05:33.230842 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-n74vh" event={"ID":"f038c921-52be-46ef-ad30-190252e42dcb","Type":"ContainerStarted","Data":"d97649f4a534cd578387ebeee2390d3bff81593747f01efc9601dd13d8217795"} Dec 05 11:05:36.629334 master-0 kubenswrapper[24928]: I1205 11:05:36.629221 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-n74vh" podStartSLOduration=9.886936005999999 podStartE2EDuration="14.629194933s" podCreationTimestamp="2025-12-05 11:05:22 +0000 UTC" firstStartedPulling="2025-12-05 11:05:27.894452429 +0000 UTC m=+1087.897646290" lastFinishedPulling="2025-12-05 11:05:32.636711366 +0000 UTC m=+1092.639905217" observedRunningTime="2025-12-05 11:05:36.597229156 +0000 UTC m=+1096.600423017" watchObservedRunningTime="2025-12-05 11:05:36.629194933 +0000 UTC m=+1096.632388794" Dec 05 11:05:37.279724 master-0 kubenswrapper[24928]: I1205 11:05:37.279648 24928 generic.go:334] "Generic (PLEG): container finished" podID="f038c921-52be-46ef-ad30-190252e42dcb" containerID="d97649f4a534cd578387ebeee2390d3bff81593747f01efc9601dd13d8217795" exitCode=0 Dec 05 11:05:37.279724 master-0 kubenswrapper[24928]: I1205 11:05:37.279706 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-n74vh" event={"ID":"f038c921-52be-46ef-ad30-190252e42dcb","Type":"ContainerDied","Data":"d97649f4a534cd578387ebeee2390d3bff81593747f01efc9601dd13d8217795"} Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.140094 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ff97bbb59-4fftj"] Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: E1205 11:05:38.140617 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e07646a-fd0f-49fc-89da-5938c26aad18" containerName="extract-content" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.140632 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e07646a-fd0f-49fc-89da-5938c26aad18" containerName="extract-content" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: E1205 11:05:38.140656 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e07646a-fd0f-49fc-89da-5938c26aad18" containerName="extract-utilities" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.140663 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e07646a-fd0f-49fc-89da-5938c26aad18" containerName="extract-utilities" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: E1205 11:05:38.140683 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f4e63e-29f0-4884-a20e-a3cb02c85235" containerName="mariadb-database-create" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.140689 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f4e63e-29f0-4884-a20e-a3cb02c85235" containerName="mariadb-database-create" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: E1205 11:05:38.140702 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c3b42bc-736b-4910-9bef-39ebeba0c5dd" containerName="mariadb-account-create-update" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.140709 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c3b42bc-736b-4910-9bef-39ebeba0c5dd" containerName="mariadb-account-create-update" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: E1205 11:05:38.140721 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d10625d-a13a-4538-9d41-6e71f1af3669" containerName="glance-db-sync" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.140726 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d10625d-a13a-4538-9d41-6e71f1af3669" containerName="glance-db-sync" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: E1205 11:05:38.140737 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1af478de-35a0-4480-a9cd-a2fa1a7ab774" containerName="mariadb-database-create" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.140743 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="1af478de-35a0-4480-a9cd-a2fa1a7ab774" containerName="mariadb-database-create" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: E1205 11:05:38.140750 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0da83455-cb84-45a3-bbe7-211be81f4b56" containerName="mariadb-account-create-update" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.140756 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0da83455-cb84-45a3-bbe7-211be81f4b56" containerName="mariadb-account-create-update" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: E1205 11:05:38.140774 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e07646a-fd0f-49fc-89da-5938c26aad18" containerName="registry-server" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.140780 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e07646a-fd0f-49fc-89da-5938c26aad18" containerName="registry-server" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.140985 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d10625d-a13a-4538-9d41-6e71f1af3669" containerName="glance-db-sync" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.141006 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f4e63e-29f0-4884-a20e-a3cb02c85235" containerName="mariadb-database-create" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.141022 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e07646a-fd0f-49fc-89da-5938c26aad18" containerName="registry-server" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.141062 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c3b42bc-736b-4910-9bef-39ebeba0c5dd" containerName="mariadb-account-create-update" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.141080 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="1af478de-35a0-4480-a9cd-a2fa1a7ab774" containerName="mariadb-database-create" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.141091 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="0da83455-cb84-45a3-bbe7-211be81f4b56" containerName="mariadb-account-create-update" Dec 05 11:05:38.146460 master-0 kubenswrapper[24928]: I1205 11:05:38.142476 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.186930 master-0 kubenswrapper[24928]: I1205 11:05:38.177980 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ff97bbb59-4fftj"] Dec 05 11:05:38.257037 master-0 kubenswrapper[24928]: I1205 11:05:38.256960 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.257275 master-0 kubenswrapper[24928]: I1205 11:05:38.257091 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plbnp\" (UniqueName: \"kubernetes.io/projected/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-kube-api-access-plbnp\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.257275 master-0 kubenswrapper[24928]: I1205 11:05:38.257151 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-dns-svc\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.257275 master-0 kubenswrapper[24928]: I1205 11:05:38.257215 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.257275 master-0 kubenswrapper[24928]: I1205 11:05:38.257243 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-config\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.322366 master-0 kubenswrapper[24928]: I1205 11:05:38.322304 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"3562e645086b700238fb78e651760aaf5a32f1ad7edaa97448b12fc0822773bd"} Dec 05 11:05:38.359328 master-0 kubenswrapper[24928]: I1205 11:05:38.359253 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.360437 master-0 kubenswrapper[24928]: I1205 11:05:38.360387 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-nb\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.360535 master-0 kubenswrapper[24928]: I1205 11:05:38.360471 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plbnp\" (UniqueName: \"kubernetes.io/projected/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-kube-api-access-plbnp\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.360621 master-0 kubenswrapper[24928]: I1205 11:05:38.360600 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-dns-svc\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.360790 master-0 kubenswrapper[24928]: I1205 11:05:38.360760 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.360855 master-0 kubenswrapper[24928]: I1205 11:05:38.360817 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-config\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.361687 master-0 kubenswrapper[24928]: I1205 11:05:38.361661 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-config\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.364139 master-0 kubenswrapper[24928]: I1205 11:05:38.362662 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-dns-svc\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.364139 master-0 kubenswrapper[24928]: I1205 11:05:38.362903 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-sb\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.402937 master-0 kubenswrapper[24928]: I1205 11:05:38.402795 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plbnp\" (UniqueName: \"kubernetes.io/projected/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-kube-api-access-plbnp\") pod \"dnsmasq-dns-5ff97bbb59-4fftj\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.534438 master-0 kubenswrapper[24928]: I1205 11:05:38.534386 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:38.766852 master-0 kubenswrapper[24928]: I1205 11:05:38.766318 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:38.877404 master-0 kubenswrapper[24928]: I1205 11:05:38.876361 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krpgx\" (UniqueName: \"kubernetes.io/projected/f038c921-52be-46ef-ad30-190252e42dcb-kube-api-access-krpgx\") pod \"f038c921-52be-46ef-ad30-190252e42dcb\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " Dec 05 11:05:38.877404 master-0 kubenswrapper[24928]: I1205 11:05:38.876504 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-combined-ca-bundle\") pod \"f038c921-52be-46ef-ad30-190252e42dcb\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " Dec 05 11:05:38.877404 master-0 kubenswrapper[24928]: I1205 11:05:38.877133 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-config-data\") pod \"f038c921-52be-46ef-ad30-190252e42dcb\" (UID: \"f038c921-52be-46ef-ad30-190252e42dcb\") " Dec 05 11:05:38.921971 master-0 kubenswrapper[24928]: I1205 11:05:38.921863 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f038c921-52be-46ef-ad30-190252e42dcb-kube-api-access-krpgx" (OuterVolumeSpecName: "kube-api-access-krpgx") pod "f038c921-52be-46ef-ad30-190252e42dcb" (UID: "f038c921-52be-46ef-ad30-190252e42dcb"). InnerVolumeSpecName "kube-api-access-krpgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:38.962665 master-0 kubenswrapper[24928]: I1205 11:05:38.962214 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f038c921-52be-46ef-ad30-190252e42dcb" (UID: "f038c921-52be-46ef-ad30-190252e42dcb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:38.980730 master-0 kubenswrapper[24928]: I1205 11:05:38.980674 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krpgx\" (UniqueName: \"kubernetes.io/projected/f038c921-52be-46ef-ad30-190252e42dcb-kube-api-access-krpgx\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:38.981044 master-0 kubenswrapper[24928]: I1205 11:05:38.981014 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:39.068722 master-0 kubenswrapper[24928]: I1205 11:05:39.068608 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-config-data" (OuterVolumeSpecName: "config-data") pod "f038c921-52be-46ef-ad30-190252e42dcb" (UID: "f038c921-52be-46ef-ad30-190252e42dcb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:39.083719 master-0 kubenswrapper[24928]: I1205 11:05:39.083083 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f038c921-52be-46ef-ad30-190252e42dcb-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:39.206437 master-0 kubenswrapper[24928]: I1205 11:05:39.206366 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ff97bbb59-4fftj"] Dec 05 11:05:39.209191 master-0 kubenswrapper[24928]: W1205 11:05:39.209100 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83a873c9_5adb_49e6_bb9c_cb9b205bff4c.slice/crio-17c5fd5933838c3d550e293e2d67953432639362d09bfbb1d1cd5a32b0f60adf WatchSource:0}: Error finding container 17c5fd5933838c3d550e293e2d67953432639362d09bfbb1d1cd5a32b0f60adf: Status 404 returned error can't find the container with id 17c5fd5933838c3d550e293e2d67953432639362d09bfbb1d1cd5a32b0f60adf Dec 05 11:05:39.337615 master-0 kubenswrapper[24928]: I1205 11:05:39.337556 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-n74vh" event={"ID":"f038c921-52be-46ef-ad30-190252e42dcb","Type":"ContainerDied","Data":"89910f1088d283452982e26d014a8925a51471c84e35192ca31ec95992bfefe8"} Dec 05 11:05:39.337615 master-0 kubenswrapper[24928]: I1205 11:05:39.337606 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89910f1088d283452982e26d014a8925a51471c84e35192ca31ec95992bfefe8" Dec 05 11:05:39.337842 master-0 kubenswrapper[24928]: I1205 11:05:39.337659 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-n74vh" Dec 05 11:05:39.341177 master-0 kubenswrapper[24928]: I1205 11:05:39.341079 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" event={"ID":"83a873c9-5adb-49e6-bb9c-cb9b205bff4c","Type":"ContainerStarted","Data":"17c5fd5933838c3d550e293e2d67953432639362d09bfbb1d1cd5a32b0f60adf"} Dec 05 11:05:39.350052 master-0 kubenswrapper[24928]: I1205 11:05:39.349989 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"d97bac1d1530277da3804d36fdcb7cab5133e18906f0433657a4ccc0955c28f8"} Dec 05 11:05:39.350052 master-0 kubenswrapper[24928]: I1205 11:05:39.350053 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"48e6d0d661fd99de5a243e222718f759c0195dc031334b791994bb5e03261541"} Dec 05 11:05:39.350312 master-0 kubenswrapper[24928]: I1205 11:05:39.350073 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"b062f5fabfe1501032a897ef3190868ef561c170f93252030d5b3121540c92b6"} Dec 05 11:05:40.058463 master-0 kubenswrapper[24928]: I1205 11:05:40.058386 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ff97bbb59-4fftj"] Dec 05 11:05:40.159512 master-0 kubenswrapper[24928]: I1205 11:05:40.151583 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-r7lkl"] Dec 05 11:05:40.159512 master-0 kubenswrapper[24928]: E1205 11:05:40.152277 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f038c921-52be-46ef-ad30-190252e42dcb" containerName="keystone-db-sync" Dec 05 11:05:40.159512 master-0 kubenswrapper[24928]: I1205 11:05:40.152300 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f038c921-52be-46ef-ad30-190252e42dcb" containerName="keystone-db-sync" Dec 05 11:05:40.159512 master-0 kubenswrapper[24928]: I1205 11:05:40.152640 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f038c921-52be-46ef-ad30-190252e42dcb" containerName="keystone-db-sync" Dec 05 11:05:40.159512 master-0 kubenswrapper[24928]: I1205 11:05:40.153536 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.159512 master-0 kubenswrapper[24928]: I1205 11:05:40.157244 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 11:05:40.159512 master-0 kubenswrapper[24928]: I1205 11:05:40.158720 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 11:05:40.159997 master-0 kubenswrapper[24928]: I1205 11:05:40.159894 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 11:05:40.166332 master-0 kubenswrapper[24928]: I1205 11:05:40.162209 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 11:05:40.379614 master-0 kubenswrapper[24928]: I1205 11:05:40.377616 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-r7lkl"] Dec 05 11:05:40.379614 master-0 kubenswrapper[24928]: I1205 11:05:40.377674 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64f947fff7-6npnx"] Dec 05 11:05:40.380244 master-0 kubenswrapper[24928]: I1205 11:05:40.379687 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:40.381797 master-0 kubenswrapper[24928]: I1205 11:05:40.380576 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.408014 master-0 kubenswrapper[24928]: I1205 11:05:40.387184 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drztt\" (UniqueName: \"kubernetes.io/projected/6140cfa3-11aa-4550-a8dc-ed81dcd087be-kube-api-access-drztt\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.408014 master-0 kubenswrapper[24928]: I1205 11:05:40.387255 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-scripts\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.408014 master-0 kubenswrapper[24928]: I1205 11:05:40.387335 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-fernet-keys\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.408014 master-0 kubenswrapper[24928]: I1205 11:05:40.387402 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-config-data\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.408014 master-0 kubenswrapper[24928]: I1205 11:05:40.388270 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-credential-keys\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.408014 master-0 kubenswrapper[24928]: I1205 11:05:40.388600 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-combined-ca-bundle\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.440334 master-0 kubenswrapper[24928]: I1205 11:05:40.440053 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64f947fff7-6npnx"] Dec 05 11:05:40.491231 master-0 kubenswrapper[24928]: I1205 11:05:40.491160 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-dns-svc\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.491485 master-0 kubenswrapper[24928]: I1205 11:05:40.491250 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-config\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.491485 master-0 kubenswrapper[24928]: I1205 11:05:40.491282 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-sb\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.491485 master-0 kubenswrapper[24928]: I1205 11:05:40.491315 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-credential-keys\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.491485 master-0 kubenswrapper[24928]: I1205 11:05:40.491360 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-nb\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.491485 master-0 kubenswrapper[24928]: I1205 11:05:40.491397 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-combined-ca-bundle\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.491485 master-0 kubenswrapper[24928]: I1205 11:05:40.491442 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cdww\" (UniqueName: \"kubernetes.io/projected/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-kube-api-access-8cdww\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.494822 master-0 kubenswrapper[24928]: I1205 11:05:40.494773 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-credential-keys\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.494971 master-0 kubenswrapper[24928]: I1205 11:05:40.494937 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drztt\" (UniqueName: \"kubernetes.io/projected/6140cfa3-11aa-4550-a8dc-ed81dcd087be-kube-api-access-drztt\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.495038 master-0 kubenswrapper[24928]: I1205 11:05:40.494990 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-scripts\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.495143 master-0 kubenswrapper[24928]: I1205 11:05:40.495110 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-fernet-keys\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.522019 master-0 kubenswrapper[24928]: I1205 11:05:40.495205 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-config-data\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.563451 master-0 kubenswrapper[24928]: I1205 11:05:40.560200 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-combined-ca-bundle\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.598227 master-0 kubenswrapper[24928]: I1205 11:05:40.597927 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-scripts\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.608036 master-0 kubenswrapper[24928]: I1205 11:05:40.606479 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"7b54939fdf4217b42263b0ef0836b8cd5593d5d018623f05a6bfa123c18c8f61"} Dec 05 11:05:40.609246 master-0 kubenswrapper[24928]: I1205 11:05:40.609197 24928 generic.go:334] "Generic (PLEG): container finished" podID="83a873c9-5adb-49e6-bb9c-cb9b205bff4c" containerID="7032e3eb858119bfbbd19ea69283e2d6bc7ef3780f4127118ea99039c71dec81" exitCode=0 Dec 05 11:05:40.609411 master-0 kubenswrapper[24928]: I1205 11:05:40.609390 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" event={"ID":"83a873c9-5adb-49e6-bb9c-cb9b205bff4c","Type":"ContainerDied","Data":"7032e3eb858119bfbbd19ea69283e2d6bc7ef3780f4127118ea99039c71dec81"} Dec 05 11:05:40.625871 master-0 kubenswrapper[24928]: I1205 11:05:40.625794 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-dns-svc\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.626604 master-0 kubenswrapper[24928]: I1205 11:05:40.625895 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-config\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.626604 master-0 kubenswrapper[24928]: I1205 11:05:40.625923 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-sb\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.626604 master-0 kubenswrapper[24928]: I1205 11:05:40.625958 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-nb\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.626604 master-0 kubenswrapper[24928]: I1205 11:05:40.626012 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cdww\" (UniqueName: \"kubernetes.io/projected/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-kube-api-access-8cdww\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.630518 master-0 kubenswrapper[24928]: I1205 11:05:40.630354 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-config\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.632925 master-0 kubenswrapper[24928]: I1205 11:05:40.632753 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-sb\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.633802 master-0 kubenswrapper[24928]: I1205 11:05:40.633137 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-nb\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.633802 master-0 kubenswrapper[24928]: I1205 11:05:40.633586 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-dns-svc\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.640872 master-0 kubenswrapper[24928]: I1205 11:05:40.640816 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-config-data\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.646269 master-0 kubenswrapper[24928]: I1205 11:05:40.646227 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drztt\" (UniqueName: \"kubernetes.io/projected/6140cfa3-11aa-4550-a8dc-ed81dcd087be-kube-api-access-drztt\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.674544 master-0 kubenswrapper[24928]: I1205 11:05:40.666247 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-fernet-keys\") pod \"keystone-bootstrap-r7lkl\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.674544 master-0 kubenswrapper[24928]: I1205 11:05:40.666876 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-db-create-x99hv"] Dec 05 11:05:40.699616 master-0 kubenswrapper[24928]: I1205 11:05:40.698108 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cdww\" (UniqueName: \"kubernetes.io/projected/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-kube-api-access-8cdww\") pod \"dnsmasq-dns-64f947fff7-6npnx\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.715791 master-0 kubenswrapper[24928]: I1205 11:05:40.715698 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-x99hv" Dec 05 11:05:40.716534 master-0 kubenswrapper[24928]: I1205 11:05:40.716466 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:40.747018 master-0 kubenswrapper[24928]: I1205 11:05:40.746963 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ec941-db-sync-9nbmc"] Dec 05 11:05:40.758530 master-0 kubenswrapper[24928]: I1205 11:05:40.755125 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.764172 master-0 kubenswrapper[24928]: I1205 11:05:40.763749 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-scripts" Dec 05 11:05:40.764172 master-0 kubenswrapper[24928]: I1205 11:05:40.764029 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-config-data" Dec 05 11:05:40.769194 master-0 kubenswrapper[24928]: I1205 11:05:40.767034 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4pf6\" (UniqueName: \"kubernetes.io/projected/415910fe-1441-44d5-a220-322e3ba3e2e3-kube-api-access-k4pf6\") pod \"ironic-db-create-x99hv\" (UID: \"415910fe-1441-44d5-a220-322e3ba3e2e3\") " pod="openstack/ironic-db-create-x99hv" Dec 05 11:05:40.769194 master-0 kubenswrapper[24928]: I1205 11:05:40.767127 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/415910fe-1441-44d5-a220-322e3ba3e2e3-operator-scripts\") pod \"ironic-db-create-x99hv\" (UID: \"415910fe-1441-44d5-a220-322e3ba3e2e3\") " pod="openstack/ironic-db-create-x99hv" Dec 05 11:05:40.800923 master-0 kubenswrapper[24928]: I1205 11:05:40.799744 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-create-x99hv"] Dec 05 11:05:40.820900 master-0 kubenswrapper[24928]: I1205 11:05:40.820841 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-db-sync-9nbmc"] Dec 05 11:05:40.834245 master-0 kubenswrapper[24928]: I1205 11:05:40.834155 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-djbqg"] Dec 05 11:05:40.840698 master-0 kubenswrapper[24928]: I1205 11:05:40.840451 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-djbqg" Dec 05 11:05:40.847299 master-0 kubenswrapper[24928]: I1205 11:05:40.844140 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 11:05:40.847299 master-0 kubenswrapper[24928]: I1205 11:05:40.844452 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 11:05:40.869070 master-0 kubenswrapper[24928]: I1205 11:05:40.867948 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-djbqg"] Dec 05 11:05:40.869197 master-0 kubenswrapper[24928]: I1205 11:05:40.869140 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-combined-ca-bundle\") pod \"neutron-db-sync-djbqg\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " pod="openstack/neutron-db-sync-djbqg" Dec 05 11:05:40.869280 master-0 kubenswrapper[24928]: I1205 11:05:40.869225 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-db-sync-config-data\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.869280 master-0 kubenswrapper[24928]: I1205 11:05:40.869258 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fdzk\" (UniqueName: \"kubernetes.io/projected/90315296-e6a2-4db3-aa3d-35af0d21a55e-kube-api-access-7fdzk\") pod \"neutron-db-sync-djbqg\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " pod="openstack/neutron-db-sync-djbqg" Dec 05 11:05:40.869380 master-0 kubenswrapper[24928]: I1205 11:05:40.869292 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/415910fe-1441-44d5-a220-322e3ba3e2e3-operator-scripts\") pod \"ironic-db-create-x99hv\" (UID: \"415910fe-1441-44d5-a220-322e3ba3e2e3\") " pod="openstack/ironic-db-create-x99hv" Dec 05 11:05:40.872077 master-0 kubenswrapper[24928]: I1205 11:05:40.870049 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/415910fe-1441-44d5-a220-322e3ba3e2e3-operator-scripts\") pod \"ironic-db-create-x99hv\" (UID: \"415910fe-1441-44d5-a220-322e3ba3e2e3\") " pod="openstack/ironic-db-create-x99hv" Dec 05 11:05:40.872077 master-0 kubenswrapper[24928]: I1205 11:05:40.870126 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-config-data\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.872077 master-0 kubenswrapper[24928]: I1205 11:05:40.870243 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68qcn\" (UniqueName: \"kubernetes.io/projected/4ae72689-6505-4064-bd26-861bda2f68cc-kube-api-access-68qcn\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.872077 master-0 kubenswrapper[24928]: I1205 11:05:40.870394 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-config\") pod \"neutron-db-sync-djbqg\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " pod="openstack/neutron-db-sync-djbqg" Dec 05 11:05:40.872077 master-0 kubenswrapper[24928]: I1205 11:05:40.870494 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ae72689-6505-4064-bd26-861bda2f68cc-etc-machine-id\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.872077 master-0 kubenswrapper[24928]: I1205 11:05:40.870648 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-scripts\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.872077 master-0 kubenswrapper[24928]: I1205 11:05:40.870676 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-combined-ca-bundle\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.872077 master-0 kubenswrapper[24928]: I1205 11:05:40.870819 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4pf6\" (UniqueName: \"kubernetes.io/projected/415910fe-1441-44d5-a220-322e3ba3e2e3-kube-api-access-k4pf6\") pod \"ironic-db-create-x99hv\" (UID: \"415910fe-1441-44d5-a220-322e3ba3e2e3\") " pod="openstack/ironic-db-create-x99hv" Dec 05 11:05:40.888521 master-0 kubenswrapper[24928]: I1205 11:05:40.885244 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-ac30-account-create-update-ngf65"] Dec 05 11:05:40.888521 master-0 kubenswrapper[24928]: I1205 11:05:40.887319 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-ac30-account-create-update-ngf65" Dec 05 11:05:40.895407 master-0 kubenswrapper[24928]: I1205 11:05:40.890207 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-db-secret" Dec 05 11:05:40.916134 master-0 kubenswrapper[24928]: I1205 11:05:40.915412 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-ac30-account-create-update-ngf65"] Dec 05 11:05:40.916134 master-0 kubenswrapper[24928]: I1205 11:05:40.915968 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4pf6\" (UniqueName: \"kubernetes.io/projected/415910fe-1441-44d5-a220-322e3ba3e2e3-kube-api-access-k4pf6\") pod \"ironic-db-create-x99hv\" (UID: \"415910fe-1441-44d5-a220-322e3ba3e2e3\") " pod="openstack/ironic-db-create-x99hv" Dec 05 11:05:40.919892 master-0 kubenswrapper[24928]: I1205 11:05:40.919646 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:40.938720 master-0 kubenswrapper[24928]: I1205 11:05:40.936256 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-rrvhc"] Dec 05 11:05:40.938720 master-0 kubenswrapper[24928]: I1205 11:05:40.938412 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:40.941689 master-0 kubenswrapper[24928]: I1205 11:05:40.941656 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 11:05:40.941895 master-0 kubenswrapper[24928]: I1205 11:05:40.941871 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 11:05:40.961995 master-0 kubenswrapper[24928]: I1205 11:05:40.961861 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-rrvhc"] Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.974295 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ae72689-6505-4064-bd26-861bda2f68cc-etc-machine-id\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.974494 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pslsc\" (UniqueName: \"kubernetes.io/projected/84942dea-64c3-403d-9d76-bbc2d9206f9f-kube-api-access-pslsc\") pod \"ironic-ac30-account-create-update-ngf65\" (UID: \"84942dea-64c3-403d-9d76-bbc2d9206f9f\") " pod="openstack/ironic-ac30-account-create-update-ngf65" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.974633 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-scripts\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.974710 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ae72689-6505-4064-bd26-861bda2f68cc-etc-machine-id\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.974773 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-config-data\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.974805 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-combined-ca-bundle\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.974947 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-combined-ca-bundle\") pod \"neutron-db-sync-djbqg\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " pod="openstack/neutron-db-sync-djbqg" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.975004 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-combined-ca-bundle\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.975043 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-db-sync-config-data\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.975109 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fdzk\" (UniqueName: \"kubernetes.io/projected/90315296-e6a2-4db3-aa3d-35af0d21a55e-kube-api-access-7fdzk\") pod \"neutron-db-sync-djbqg\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " pod="openstack/neutron-db-sync-djbqg" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.975156 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-scripts\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.975191 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-config-data\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.975340 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68qcn\" (UniqueName: \"kubernetes.io/projected/4ae72689-6505-4064-bd26-861bda2f68cc-kube-api-access-68qcn\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.975388 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dp4q\" (UniqueName: \"kubernetes.io/projected/cd047ad1-2e39-47f6-ad8f-97eb080e7766-kube-api-access-5dp4q\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.975729 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84942dea-64c3-403d-9d76-bbc2d9206f9f-operator-scripts\") pod \"ironic-ac30-account-create-update-ngf65\" (UID: \"84942dea-64c3-403d-9d76-bbc2d9206f9f\") " pod="openstack/ironic-ac30-account-create-update-ngf65" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.975784 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd047ad1-2e39-47f6-ad8f-97eb080e7766-logs\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:40.976795 master-0 kubenswrapper[24928]: I1205 11:05:40.976488 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-config\") pod \"neutron-db-sync-djbqg\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " pod="openstack/neutron-db-sync-djbqg" Dec 05 11:05:40.980656 master-0 kubenswrapper[24928]: I1205 11:05:40.980607 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-db-sync-config-data\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:40.984770 master-0 kubenswrapper[24928]: I1205 11:05:40.981231 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64f947fff7-6npnx"] Dec 05 11:05:40.992333 master-0 kubenswrapper[24928]: I1205 11:05:40.992283 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-config\") pod \"neutron-db-sync-djbqg\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " pod="openstack/neutron-db-sync-djbqg" Dec 05 11:05:41.000734 master-0 kubenswrapper[24928]: I1205 11:05:40.999888 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x5nq4"] Dec 05 11:05:41.017282 master-0 kubenswrapper[24928]: I1205 11:05:41.017225 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-scripts\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:41.017483 master-0 kubenswrapper[24928]: I1205 11:05:41.017438 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-combined-ca-bundle\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:41.017542 master-0 kubenswrapper[24928]: I1205 11:05:41.017465 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-combined-ca-bundle\") pod \"neutron-db-sync-djbqg\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " pod="openstack/neutron-db-sync-djbqg" Dec 05 11:05:41.018642 master-0 kubenswrapper[24928]: I1205 11:05:41.018084 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-config-data\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:41.021997 master-0 kubenswrapper[24928]: I1205 11:05:41.021937 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68qcn\" (UniqueName: \"kubernetes.io/projected/4ae72689-6505-4064-bd26-861bda2f68cc-kube-api-access-68qcn\") pod \"cinder-ec941-db-sync-9nbmc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:41.044019 master-0 kubenswrapper[24928]: I1205 11:05:41.043913 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fdzk\" (UniqueName: \"kubernetes.io/projected/90315296-e6a2-4db3-aa3d-35af0d21a55e-kube-api-access-7fdzk\") pod \"neutron-db-sync-djbqg\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " pod="openstack/neutron-db-sync-djbqg" Dec 05 11:05:41.044221 master-0 kubenswrapper[24928]: I1205 11:05:41.044020 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cbb8c9bdf-4rkl7"] Dec 05 11:05:41.046654 master-0 kubenswrapper[24928]: I1205 11:05:41.046598 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.053546 master-0 kubenswrapper[24928]: I1205 11:05:41.053092 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cbb8c9bdf-4rkl7"] Dec 05 11:05:41.107767 master-0 kubenswrapper[24928]: I1205 11:05:41.105852 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:05:41.107767 master-0 kubenswrapper[24928]: I1205 11:05:41.106909 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-x99hv" Dec 05 11:05:41.107767 master-0 kubenswrapper[24928]: I1205 11:05:41.107728 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-dns-svc\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.109011 master-0 kubenswrapper[24928]: I1205 11:05:41.107793 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wz5q\" (UniqueName: \"kubernetes.io/projected/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-kube-api-access-2wz5q\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.109011 master-0 kubenswrapper[24928]: I1205 11:05:41.107860 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-nb\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.109011 master-0 kubenswrapper[24928]: I1205 11:05:41.107898 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dp4q\" (UniqueName: \"kubernetes.io/projected/cd047ad1-2e39-47f6-ad8f-97eb080e7766-kube-api-access-5dp4q\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:41.109011 master-0 kubenswrapper[24928]: I1205 11:05:41.107950 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84942dea-64c3-403d-9d76-bbc2d9206f9f-operator-scripts\") pod \"ironic-ac30-account-create-update-ngf65\" (UID: \"84942dea-64c3-403d-9d76-bbc2d9206f9f\") " pod="openstack/ironic-ac30-account-create-update-ngf65" Dec 05 11:05:41.109011 master-0 kubenswrapper[24928]: I1205 11:05:41.107987 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-sb\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.109011 master-0 kubenswrapper[24928]: I1205 11:05:41.108019 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd047ad1-2e39-47f6-ad8f-97eb080e7766-logs\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:41.109011 master-0 kubenswrapper[24928]: I1205 11:05:41.108087 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pslsc\" (UniqueName: \"kubernetes.io/projected/84942dea-64c3-403d-9d76-bbc2d9206f9f-kube-api-access-pslsc\") pod \"ironic-ac30-account-create-update-ngf65\" (UID: \"84942dea-64c3-403d-9d76-bbc2d9206f9f\") " pod="openstack/ironic-ac30-account-create-update-ngf65" Dec 05 11:05:41.109011 master-0 kubenswrapper[24928]: I1205 11:05:41.108792 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd047ad1-2e39-47f6-ad8f-97eb080e7766-logs\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:41.109805 master-0 kubenswrapper[24928]: I1205 11:05:41.109695 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84942dea-64c3-403d-9d76-bbc2d9206f9f-operator-scripts\") pod \"ironic-ac30-account-create-update-ngf65\" (UID: \"84942dea-64c3-403d-9d76-bbc2d9206f9f\") " pod="openstack/ironic-ac30-account-create-update-ngf65" Dec 05 11:05:41.109805 master-0 kubenswrapper[24928]: I1205 11:05:41.109768 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-config-data\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:41.109982 master-0 kubenswrapper[24928]: I1205 11:05:41.109937 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-combined-ca-bundle\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:41.110043 master-0 kubenswrapper[24928]: I1205 11:05:41.110019 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-scripts\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:41.112547 master-0 kubenswrapper[24928]: I1205 11:05:41.111783 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-config\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.117386 master-0 kubenswrapper[24928]: I1205 11:05:41.117345 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-config-data\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:41.120286 master-0 kubenswrapper[24928]: I1205 11:05:41.120221 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-combined-ca-bundle\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:41.120445 master-0 kubenswrapper[24928]: I1205 11:05:41.120331 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-scripts\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:41.135717 master-0 kubenswrapper[24928]: I1205 11:05:41.135663 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pslsc\" (UniqueName: \"kubernetes.io/projected/84942dea-64c3-403d-9d76-bbc2d9206f9f-kube-api-access-pslsc\") pod \"ironic-ac30-account-create-update-ngf65\" (UID: \"84942dea-64c3-403d-9d76-bbc2d9206f9f\") " pod="openstack/ironic-ac30-account-create-update-ngf65" Dec 05 11:05:41.136172 master-0 kubenswrapper[24928]: I1205 11:05:41.136126 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dp4q\" (UniqueName: \"kubernetes.io/projected/cd047ad1-2e39-47f6-ad8f-97eb080e7766-kube-api-access-5dp4q\") pod \"placement-db-sync-rrvhc\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:41.219776 master-0 kubenswrapper[24928]: I1205 11:05:41.219609 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-config\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.220587 master-0 kubenswrapper[24928]: I1205 11:05:41.220519 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-dns-svc\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.220587 master-0 kubenswrapper[24928]: I1205 11:05:41.220564 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wz5q\" (UniqueName: \"kubernetes.io/projected/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-kube-api-access-2wz5q\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.220587 master-0 kubenswrapper[24928]: I1205 11:05:41.220585 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-nb\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.220863 master-0 kubenswrapper[24928]: I1205 11:05:41.220659 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-sb\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.221774 master-0 kubenswrapper[24928]: I1205 11:05:41.221746 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-sb\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.224759 master-0 kubenswrapper[24928]: I1205 11:05:41.224518 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-dns-svc\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.225291 master-0 kubenswrapper[24928]: I1205 11:05:41.225258 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-config\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.225816 master-0 kubenswrapper[24928]: I1205 11:05:41.225783 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-djbqg" Dec 05 11:05:41.227064 master-0 kubenswrapper[24928]: I1205 11:05:41.226909 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-nb\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.229713 master-0 kubenswrapper[24928]: I1205 11:05:41.229667 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-ac30-account-create-update-ngf65" Dec 05 11:05:41.246433 master-0 kubenswrapper[24928]: I1205 11:05:41.246371 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wz5q\" (UniqueName: \"kubernetes.io/projected/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-kube-api-access-2wz5q\") pod \"dnsmasq-dns-cbb8c9bdf-4rkl7\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.438335 master-0 kubenswrapper[24928]: I1205 11:05:41.433496 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vdmr2" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerName="registry-server" probeResult="failure" output=< Dec 05 11:05:41.438335 master-0 kubenswrapper[24928]: timeout: failed to connect service ":50051" within 1s Dec 05 11:05:41.438335 master-0 kubenswrapper[24928]: > Dec 05 11:05:41.484582 master-0 kubenswrapper[24928]: I1205 11:05:41.484319 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-rrvhc" Dec 05 11:05:41.545455 master-0 kubenswrapper[24928]: I1205 11:05:41.543550 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:41.649450 master-0 kubenswrapper[24928]: I1205 11:05:41.646647 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64f947fff7-6npnx"] Dec 05 11:05:41.665924 master-0 kubenswrapper[24928]: I1205 11:05:41.665746 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"c33ef0b093cd4df8519191c96ca5bbd42f020ec98cc29d2697c8dd7b75a3076f"} Dec 05 11:05:41.665924 master-0 kubenswrapper[24928]: I1205 11:05:41.665926 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"cc6a7a43-f73c-4e31-a61e-de8f4a5957b6","Type":"ContainerStarted","Data":"7fff17dbcf87d9b4eee31cd213cd8ff4831d31c3396429bcd7d02579549d3ad9"} Dec 05 11:05:41.666190 master-0 kubenswrapper[24928]: I1205 11:05:41.665890 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x5nq4" podUID="032f0939-20e7-46cb-9665-b1346b7e1b10" containerName="registry-server" containerID="cri-o://1ad8d4221e10a10b242cfb51bce3c73c08c96b11c2a121dce15f2c8ce277ce37" gracePeriod=2 Dec 05 11:05:41.688178 master-0 kubenswrapper[24928]: E1205 11:05:41.688109 24928 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 05 11:05:41.688178 master-0 kubenswrapper[24928]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/83a873c9-5adb-49e6-bb9c-cb9b205bff4c/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 05 11:05:41.688178 master-0 kubenswrapper[24928]: > podSandboxID="17c5fd5933838c3d550e293e2d67953432639362d09bfbb1d1cd5a32b0f60adf" Dec 05 11:05:41.688474 master-0 kubenswrapper[24928]: E1205 11:05:41.688322 24928 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 05 11:05:41.688474 master-0 kubenswrapper[24928]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nbfh5b4hd6h66dh55h5c6h95hc4h584h8bhc7h648h64chd4h54chbbh5b7h654hdh9bhcch5f9h64chcfh65h95h55fhbh685hfdh64fh596q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-plbnp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000800000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ff97bbb59-4fftj_openstack(83a873c9-5adb-49e6-bb9c-cb9b205bff4c): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/83a873c9-5adb-49e6-bb9c-cb9b205bff4c/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 05 11:05:41.688474 master-0 kubenswrapper[24928]: > logger="UnhandledError" Dec 05 11:05:41.689518 master-0 kubenswrapper[24928]: E1205 11:05:41.689446 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/83a873c9-5adb-49e6-bb9c-cb9b205bff4c/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" podUID="83a873c9-5adb-49e6-bb9c-cb9b205bff4c" Dec 05 11:05:41.697591 master-0 kubenswrapper[24928]: W1205 11:05:41.697538 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bcd1d9a_8ff3_4577_8308_bd63c025d4e4.slice/crio-bc3dc9f5b31d8d3a20a93174bc372f83eddfe539bb50d5c3be8a716b732ea96b WatchSource:0}: Error finding container bc3dc9f5b31d8d3a20a93174bc372f83eddfe539bb50d5c3be8a716b732ea96b: Status 404 returned error can't find the container with id bc3dc9f5b31d8d3a20a93174bc372f83eddfe539bb50d5c3be8a716b732ea96b Dec 05 11:05:41.770327 master-0 kubenswrapper[24928]: I1205 11:05:41.769857 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=23.697267734 podStartE2EDuration="55.769832496s" podCreationTimestamp="2025-12-05 11:04:46 +0000 UTC" firstStartedPulling="2025-12-05 11:05:05.794218707 +0000 UTC m=+1065.797412558" lastFinishedPulling="2025-12-05 11:05:37.866783459 +0000 UTC m=+1097.869977320" observedRunningTime="2025-12-05 11:05:41.75114211 +0000 UTC m=+1101.754335961" watchObservedRunningTime="2025-12-05 11:05:41.769832496 +0000 UTC m=+1101.773026347" Dec 05 11:05:41.897834 master-0 kubenswrapper[24928]: W1205 11:05:41.897692 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ae72689_6505_4064_bd26_861bda2f68cc.slice/crio-612bf794def52a95657619b5c41bbfc0aa5e330095d97560da0f18adc3833336 WatchSource:0}: Error finding container 612bf794def52a95657619b5c41bbfc0aa5e330095d97560da0f18adc3833336: Status 404 returned error can't find the container with id 612bf794def52a95657619b5c41bbfc0aa5e330095d97560da0f18adc3833336 Dec 05 11:05:41.943727 master-0 kubenswrapper[24928]: I1205 11:05:41.943543 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-db-sync-9nbmc"] Dec 05 11:05:41.983462 master-0 kubenswrapper[24928]: I1205 11:05:41.972859 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-r7lkl"] Dec 05 11:05:42.236441 master-0 kubenswrapper[24928]: I1205 11:05:42.235939 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cbb8c9bdf-4rkl7"] Dec 05 11:05:42.249491 master-0 kubenswrapper[24928]: I1205 11:05:42.248484 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688fbc9bcc-96d2d"] Dec 05 11:05:42.266056 master-0 kubenswrapper[24928]: I1205 11:05:42.265943 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.275279 master-0 kubenswrapper[24928]: I1205 11:05:42.272027 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 05 11:05:42.325556 master-0 kubenswrapper[24928]: I1205 11:05:42.322834 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688fbc9bcc-96d2d"] Dec 05 11:05:42.354402 master-0 kubenswrapper[24928]: I1205 11:05:42.353154 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:05:42.360536 master-0 kubenswrapper[24928]: I1205 11:05:42.356762 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.360536 master-0 kubenswrapper[24928]: I1205 11:05:42.359845 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-ec941-default-external-config-data" Dec 05 11:05:42.360536 master-0 kubenswrapper[24928]: I1205 11:05:42.360434 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 05 11:05:42.375379 master-0 kubenswrapper[24928]: I1205 11:05:42.375293 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:05:42.381682 master-0 kubenswrapper[24928]: I1205 11:05:42.381553 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-nb\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.381851 master-0 kubenswrapper[24928]: I1205 11:05:42.381769 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-svc\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.381851 master-0 kubenswrapper[24928]: I1205 11:05:42.381797 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-config\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.381917 master-0 kubenswrapper[24928]: I1205 11:05:42.381868 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-sb\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.381917 master-0 kubenswrapper[24928]: I1205 11:05:42.381908 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-swift-storage-0\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.381981 master-0 kubenswrapper[24928]: I1205 11:05:42.381949 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxbth\" (UniqueName: \"kubernetes.io/projected/628a74ee-1afa-4e41-8b19-c6b0402a3368-kube-api-access-vxbth\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.397054 master-0 kubenswrapper[24928]: I1205 11:05:42.396980 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-ac30-account-create-update-ngf65"] Dec 05 11:05:42.406180 master-0 kubenswrapper[24928]: I1205 11:05:42.406077 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-create-x99hv"] Dec 05 11:05:42.428452 master-0 kubenswrapper[24928]: W1205 11:05:42.428355 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod84942dea_64c3_403d_9d76_bbc2d9206f9f.slice/crio-949177e2d5f42a0e3bbcecac140fa1c4deadacef7f1617d32fe60dcb86112cf6 WatchSource:0}: Error finding container 949177e2d5f42a0e3bbcecac140fa1c4deadacef7f1617d32fe60dcb86112cf6: Status 404 returned error can't find the container with id 949177e2d5f42a0e3bbcecac140fa1c4deadacef7f1617d32fe60dcb86112cf6 Dec 05 11:05:42.434375 master-0 kubenswrapper[24928]: W1205 11:05:42.433581 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod415910fe_1441_44d5_a220_322e3ba3e2e3.slice/crio-021e7f11b0e27fa74a4cb3e7e4e907a609e13912b8ae022d6031e3fc609c6a9f WatchSource:0}: Error finding container 021e7f11b0e27fa74a4cb3e7e4e907a609e13912b8ae022d6031e3fc609c6a9f: Status 404 returned error can't find the container with id 021e7f11b0e27fa74a4cb3e7e4e907a609e13912b8ae022d6031e3fc609c6a9f Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489116 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-nb\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489232 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489294 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-combined-ca-bundle\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489331 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-scripts\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489384 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-logs\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489415 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-config\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489453 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-svc\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489515 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-httpd-run\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489549 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-sb\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489582 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-swift-storage-0\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489624 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxbth\" (UniqueName: \"kubernetes.io/projected/628a74ee-1afa-4e41-8b19-c6b0402a3368-kube-api-access-vxbth\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489668 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-config-data\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.489697 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcm2z\" (UniqueName: \"kubernetes.io/projected/577d038e-b40d-43e5-a6ad-ad4e278a3213-kube-api-access-vcm2z\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.490805 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-nb\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.491855 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-config\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.492480 master-0 kubenswrapper[24928]: I1205 11:05:42.492480 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-svc\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.493454 master-0 kubenswrapper[24928]: I1205 11:05:42.493150 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-sb\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.496178 master-0 kubenswrapper[24928]: I1205 11:05:42.495143 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-swift-storage-0\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.520205 master-0 kubenswrapper[24928]: I1205 11:05:42.520094 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxbth\" (UniqueName: \"kubernetes.io/projected/628a74ee-1afa-4e41-8b19-c6b0402a3368-kube-api-access-vxbth\") pod \"dnsmasq-dns-688fbc9bcc-96d2d\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.598343 master-0 kubenswrapper[24928]: I1205 11:05:42.598263 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.598343 master-0 kubenswrapper[24928]: I1205 11:05:42.598350 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-combined-ca-bundle\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.598682 master-0 kubenswrapper[24928]: I1205 11:05:42.598389 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-scripts\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.598682 master-0 kubenswrapper[24928]: I1205 11:05:42.598482 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-logs\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.598682 master-0 kubenswrapper[24928]: I1205 11:05:42.598582 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-httpd-run\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.598803 master-0 kubenswrapper[24928]: I1205 11:05:42.598718 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-config-data\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.606367 master-0 kubenswrapper[24928]: I1205 11:05:42.598745 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcm2z\" (UniqueName: \"kubernetes.io/projected/577d038e-b40d-43e5-a6ad-ad4e278a3213-kube-api-access-vcm2z\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.608778 master-0 kubenswrapper[24928]: I1205 11:05:42.607560 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-config-data\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.609393 master-0 kubenswrapper[24928]: I1205 11:05:42.601612 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-logs\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.609523 master-0 kubenswrapper[24928]: I1205 11:05:42.605348 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-scripts\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.609749 master-0 kubenswrapper[24928]: I1205 11:05:42.606026 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:05:42.611041 master-0 kubenswrapper[24928]: I1205 11:05:42.610987 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/4498c228462d6dd9d5d70d1a4aafdb52e778f9b7d49e65869f766f31c3adb188/globalmount\"" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.611490 master-0 kubenswrapper[24928]: I1205 11:05:42.601568 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-httpd-run\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.625724 master-0 kubenswrapper[24928]: I1205 11:05:42.625675 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-combined-ca-bundle\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.632659 master-0 kubenswrapper[24928]: I1205 11:05:42.632606 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcm2z\" (UniqueName: \"kubernetes.io/projected/577d038e-b40d-43e5-a6ad-ad4e278a3213-kube-api-access-vcm2z\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:42.717366 master-0 kubenswrapper[24928]: I1205 11:05:42.717301 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-db-sync-9nbmc" event={"ID":"4ae72689-6505-4064-bd26-861bda2f68cc","Type":"ContainerStarted","Data":"612bf794def52a95657619b5c41bbfc0aa5e330095d97560da0f18adc3833336"} Dec 05 11:05:42.719953 master-0 kubenswrapper[24928]: I1205 11:05:42.719921 24928 generic.go:334] "Generic (PLEG): container finished" podID="4bcd1d9a-8ff3-4577-8308-bd63c025d4e4" containerID="8642da1100670801b9eb83c9f0a447ca91642e8d3b1a211a1f3463a99d202cd1" exitCode=0 Dec 05 11:05:42.720616 master-0 kubenswrapper[24928]: I1205 11:05:42.720595 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f947fff7-6npnx" event={"ID":"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4","Type":"ContainerDied","Data":"8642da1100670801b9eb83c9f0a447ca91642e8d3b1a211a1f3463a99d202cd1"} Dec 05 11:05:42.720804 master-0 kubenswrapper[24928]: I1205 11:05:42.720770 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f947fff7-6npnx" event={"ID":"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4","Type":"ContainerStarted","Data":"bc3dc9f5b31d8d3a20a93174bc372f83eddfe539bb50d5c3be8a716b732ea96b"} Dec 05 11:05:42.744205 master-0 kubenswrapper[24928]: I1205 11:05:42.744149 24928 generic.go:334] "Generic (PLEG): container finished" podID="032f0939-20e7-46cb-9665-b1346b7e1b10" containerID="1ad8d4221e10a10b242cfb51bce3c73c08c96b11c2a121dce15f2c8ce277ce37" exitCode=0 Dec 05 11:05:42.744794 master-0 kubenswrapper[24928]: I1205 11:05:42.744764 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5nq4" event={"ID":"032f0939-20e7-46cb-9665-b1346b7e1b10","Type":"ContainerDied","Data":"1ad8d4221e10a10b242cfb51bce3c73c08c96b11c2a121dce15f2c8ce277ce37"} Dec 05 11:05:42.758761 master-0 kubenswrapper[24928]: I1205 11:05:42.758704 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-ac30-account-create-update-ngf65" event={"ID":"84942dea-64c3-403d-9d76-bbc2d9206f9f","Type":"ContainerStarted","Data":"949177e2d5f42a0e3bbcecac140fa1c4deadacef7f1617d32fe60dcb86112cf6"} Dec 05 11:05:42.774844 master-0 kubenswrapper[24928]: I1205 11:05:42.773527 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-x99hv" event={"ID":"415910fe-1441-44d5-a220-322e3ba3e2e3","Type":"ContainerStarted","Data":"021e7f11b0e27fa74a4cb3e7e4e907a609e13912b8ae022d6031e3fc609c6a9f"} Dec 05 11:05:42.775618 master-0 kubenswrapper[24928]: I1205 11:05:42.775581 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:42.910547 master-0 kubenswrapper[24928]: I1205 11:05:42.849988 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r7lkl" event={"ID":"6140cfa3-11aa-4550-a8dc-ed81dcd087be","Type":"ContainerStarted","Data":"147d4e2b823107dd73c2695e1569f383bc3f459b9c676b6a321be91f112e4b3e"} Dec 05 11:05:42.910547 master-0 kubenswrapper[24928]: I1205 11:05:42.850098 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r7lkl" event={"ID":"6140cfa3-11aa-4550-a8dc-ed81dcd087be","Type":"ContainerStarted","Data":"95e5cfae88fd5e3131dff9262f4a9817a8cc318f85370c21381382497a509fbd"} Dec 05 11:05:42.910547 master-0 kubenswrapper[24928]: I1205 11:05:42.859143 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-djbqg"] Dec 05 11:05:42.910547 master-0 kubenswrapper[24928]: I1205 11:05:42.900255 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-rrvhc"] Dec 05 11:05:42.910547 master-0 kubenswrapper[24928]: I1205 11:05:42.900330 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cbb8c9bdf-4rkl7"] Dec 05 11:05:42.967452 master-0 kubenswrapper[24928]: I1205 11:05:42.964240 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-r7lkl" podStartSLOduration=2.964215045 podStartE2EDuration="2.964215045s" podCreationTimestamp="2025-12-05 11:05:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:05:42.943520199 +0000 UTC m=+1102.946714070" watchObservedRunningTime="2025-12-05 11:05:42.964215045 +0000 UTC m=+1102.967408916" Dec 05 11:05:43.015814 master-0 kubenswrapper[24928]: W1205 11:05:43.013701 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cd6a2cb_720d_4653_bb1a_9b102506d7d3.slice/crio-0b81c9b78b31c639fe1c47540edced7d4d0add5fdfb03726fc863e0ea2fe5ef0 WatchSource:0}: Error finding container 0b81c9b78b31c639fe1c47540edced7d4d0add5fdfb03726fc863e0ea2fe5ef0: Status 404 returned error can't find the container with id 0b81c9b78b31c639fe1c47540edced7d4d0add5fdfb03726fc863e0ea2fe5ef0 Dec 05 11:05:43.044600 master-0 kubenswrapper[24928]: I1205 11:05:43.044081 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:43.165896 master-0 kubenswrapper[24928]: I1205 11:05:43.161159 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9z8g\" (UniqueName: \"kubernetes.io/projected/032f0939-20e7-46cb-9665-b1346b7e1b10-kube-api-access-x9z8g\") pod \"032f0939-20e7-46cb-9665-b1346b7e1b10\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " Dec 05 11:05:43.165896 master-0 kubenswrapper[24928]: I1205 11:05:43.161318 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-catalog-content\") pod \"032f0939-20e7-46cb-9665-b1346b7e1b10\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " Dec 05 11:05:43.165896 master-0 kubenswrapper[24928]: I1205 11:05:43.162885 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-utilities\") pod \"032f0939-20e7-46cb-9665-b1346b7e1b10\" (UID: \"032f0939-20e7-46cb-9665-b1346b7e1b10\") " Dec 05 11:05:43.165896 master-0 kubenswrapper[24928]: I1205 11:05:43.164958 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-utilities" (OuterVolumeSpecName: "utilities") pod "032f0939-20e7-46cb-9665-b1346b7e1b10" (UID: "032f0939-20e7-46cb-9665-b1346b7e1b10"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:05:43.172585 master-0 kubenswrapper[24928]: I1205 11:05:43.168549 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/032f0939-20e7-46cb-9665-b1346b7e1b10-kube-api-access-x9z8g" (OuterVolumeSpecName: "kube-api-access-x9z8g") pod "032f0939-20e7-46cb-9665-b1346b7e1b10" (UID: "032f0939-20e7-46cb-9665-b1346b7e1b10"). InnerVolumeSpecName "kube-api-access-x9z8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:43.267726 master-0 kubenswrapper[24928]: I1205 11:05:43.266786 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9z8g\" (UniqueName: \"kubernetes.io/projected/032f0939-20e7-46cb-9665-b1346b7e1b10-kube-api-access-x9z8g\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:43.267726 master-0 kubenswrapper[24928]: I1205 11:05:43.266848 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:43.430381 master-0 kubenswrapper[24928]: I1205 11:05:43.429487 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "032f0939-20e7-46cb-9665-b1346b7e1b10" (UID: "032f0939-20e7-46cb-9665-b1346b7e1b10"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:05:43.472494 master-0 kubenswrapper[24928]: I1205 11:05:43.472278 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:05:43.478895 master-0 kubenswrapper[24928]: I1205 11:05:43.478818 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/032f0939-20e7-46cb-9665-b1346b7e1b10-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:43.479045 master-0 kubenswrapper[24928]: E1205 11:05:43.478958 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032f0939-20e7-46cb-9665-b1346b7e1b10" containerName="extract-utilities" Dec 05 11:05:43.479045 master-0 kubenswrapper[24928]: I1205 11:05:43.478979 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="032f0939-20e7-46cb-9665-b1346b7e1b10" containerName="extract-utilities" Dec 05 11:05:43.479045 master-0 kubenswrapper[24928]: E1205 11:05:43.479024 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032f0939-20e7-46cb-9665-b1346b7e1b10" containerName="extract-content" Dec 05 11:05:43.479045 master-0 kubenswrapper[24928]: I1205 11:05:43.479036 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="032f0939-20e7-46cb-9665-b1346b7e1b10" containerName="extract-content" Dec 05 11:05:43.482230 master-0 kubenswrapper[24928]: E1205 11:05:43.479503 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="032f0939-20e7-46cb-9665-b1346b7e1b10" containerName="registry-server" Dec 05 11:05:43.482230 master-0 kubenswrapper[24928]: I1205 11:05:43.479521 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="032f0939-20e7-46cb-9665-b1346b7e1b10" containerName="registry-server" Dec 05 11:05:43.482230 master-0 kubenswrapper[24928]: I1205 11:05:43.480340 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="032f0939-20e7-46cb-9665-b1346b7e1b10" containerName="registry-server" Dec 05 11:05:43.487458 master-0 kubenswrapper[24928]: I1205 11:05:43.487355 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.489639 master-0 kubenswrapper[24928]: I1205 11:05:43.489551 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:05:43.505098 master-0 kubenswrapper[24928]: I1205 11:05:43.493553 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-ec941-default-internal-config-data" Dec 05 11:05:43.585195 master-0 kubenswrapper[24928]: I1205 11:05:43.581498 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.585195 master-0 kubenswrapper[24928]: I1205 11:05:43.581660 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.585195 master-0 kubenswrapper[24928]: I1205 11:05:43.581773 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.585195 master-0 kubenswrapper[24928]: I1205 11:05:43.582137 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.585195 master-0 kubenswrapper[24928]: I1205 11:05:43.582230 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.585195 master-0 kubenswrapper[24928]: I1205 11:05:43.582272 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rb8r\" (UniqueName: \"kubernetes.io/projected/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-kube-api-access-7rb8r\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.585195 master-0 kubenswrapper[24928]: I1205 11:05:43.582352 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.590609 master-0 kubenswrapper[24928]: I1205 11:05:43.590244 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:43.684627 master-0 kubenswrapper[24928]: I1205 11:05:43.683471 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-dns-svc\") pod \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " Dec 05 11:05:43.684627 master-0 kubenswrapper[24928]: I1205 11:05:43.683580 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-sb\") pod \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " Dec 05 11:05:43.684627 master-0 kubenswrapper[24928]: I1205 11:05:43.683625 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-config\") pod \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " Dec 05 11:05:43.684627 master-0 kubenswrapper[24928]: I1205 11:05:43.683655 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cdww\" (UniqueName: \"kubernetes.io/projected/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-kube-api-access-8cdww\") pod \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " Dec 05 11:05:43.684627 master-0 kubenswrapper[24928]: I1205 11:05:43.683692 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-nb\") pod \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\" (UID: \"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4\") " Dec 05 11:05:43.684627 master-0 kubenswrapper[24928]: I1205 11:05:43.684130 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.684627 master-0 kubenswrapper[24928]: I1205 11:05:43.684207 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.684627 master-0 kubenswrapper[24928]: I1205 11:05:43.684264 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.688281 master-0 kubenswrapper[24928]: I1205 11:05:43.688238 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.688477 master-0 kubenswrapper[24928]: I1205 11:05:43.688461 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.688571 master-0 kubenswrapper[24928]: I1205 11:05:43.688555 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rb8r\" (UniqueName: \"kubernetes.io/projected/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-kube-api-access-7rb8r\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.688736 master-0 kubenswrapper[24928]: I1205 11:05:43.688722 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.701501 master-0 kubenswrapper[24928]: I1205 11:05:43.692875 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-kube-api-access-8cdww" (OuterVolumeSpecName: "kube-api-access-8cdww") pod "4bcd1d9a-8ff3-4577-8308-bd63c025d4e4" (UID: "4bcd1d9a-8ff3-4577-8308-bd63c025d4e4"). InnerVolumeSpecName "kube-api-access-8cdww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:43.701501 master-0 kubenswrapper[24928]: I1205 11:05:43.693797 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.701501 master-0 kubenswrapper[24928]: I1205 11:05:43.696199 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.707740 master-0 kubenswrapper[24928]: I1205 11:05:43.706386 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.707740 master-0 kubenswrapper[24928]: I1205 11:05:43.706540 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.707740 master-0 kubenswrapper[24928]: I1205 11:05:43.706982 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:05:43.707740 master-0 kubenswrapper[24928]: I1205 11:05:43.707031 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/f829e9071eee306944ccf41eabbcf74a1473d7031ba724c821f61b9a30a289e9/globalmount\"" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.714216 master-0 kubenswrapper[24928]: I1205 11:05:43.714146 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.732514 master-0 kubenswrapper[24928]: I1205 11:05:43.732470 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:43.755123 master-0 kubenswrapper[24928]: I1205 11:05:43.755068 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-config" (OuterVolumeSpecName: "config") pod "4bcd1d9a-8ff3-4577-8308-bd63c025d4e4" (UID: "4bcd1d9a-8ff3-4577-8308-bd63c025d4e4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:43.766541 master-0 kubenswrapper[24928]: I1205 11:05:43.766323 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rb8r\" (UniqueName: \"kubernetes.io/projected/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-kube-api-access-7rb8r\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:43.783391 master-0 kubenswrapper[24928]: I1205 11:05:43.783316 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4bcd1d9a-8ff3-4577-8308-bd63c025d4e4" (UID: "4bcd1d9a-8ff3-4577-8308-bd63c025d4e4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:43.793893 master-0 kubenswrapper[24928]: I1205 11:05:43.792720 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-config\") pod \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " Dec 05 11:05:43.793893 master-0 kubenswrapper[24928]: I1205 11:05:43.792945 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-sb\") pod \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " Dec 05 11:05:43.793893 master-0 kubenswrapper[24928]: I1205 11:05:43.793031 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-dns-svc\") pod \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " Dec 05 11:05:43.793893 master-0 kubenswrapper[24928]: I1205 11:05:43.793089 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plbnp\" (UniqueName: \"kubernetes.io/projected/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-kube-api-access-plbnp\") pod \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " Dec 05 11:05:43.793893 master-0 kubenswrapper[24928]: I1205 11:05:43.793153 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-nb\") pod \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\" (UID: \"83a873c9-5adb-49e6-bb9c-cb9b205bff4c\") " Dec 05 11:05:43.793893 master-0 kubenswrapper[24928]: I1205 11:05:43.793769 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-sb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:43.793893 master-0 kubenswrapper[24928]: I1205 11:05:43.793795 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:43.793893 master-0 kubenswrapper[24928]: I1205 11:05:43.793813 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cdww\" (UniqueName: \"kubernetes.io/projected/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-kube-api-access-8cdww\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:43.800115 master-0 kubenswrapper[24928]: I1205 11:05:43.800054 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4bcd1d9a-8ff3-4577-8308-bd63c025d4e4" (UID: "4bcd1d9a-8ff3-4577-8308-bd63c025d4e4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:43.802504 master-0 kubenswrapper[24928]: I1205 11:05:43.802374 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-kube-api-access-plbnp" (OuterVolumeSpecName: "kube-api-access-plbnp") pod "83a873c9-5adb-49e6-bb9c-cb9b205bff4c" (UID: "83a873c9-5adb-49e6-bb9c-cb9b205bff4c"). InnerVolumeSpecName "kube-api-access-plbnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:43.855085 master-0 kubenswrapper[24928]: I1205 11:05:43.851187 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4bcd1d9a-8ff3-4577-8308-bd63c025d4e4" (UID: "4bcd1d9a-8ff3-4577-8308-bd63c025d4e4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:43.855085 master-0 kubenswrapper[24928]: I1205 11:05:43.854943 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "83a873c9-5adb-49e6-bb9c-cb9b205bff4c" (UID: "83a873c9-5adb-49e6-bb9c-cb9b205bff4c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:43.868126 master-0 kubenswrapper[24928]: I1205 11:05:43.868060 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "83a873c9-5adb-49e6-bb9c-cb9b205bff4c" (UID: "83a873c9-5adb-49e6-bb9c-cb9b205bff4c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:43.876171 master-0 kubenswrapper[24928]: I1205 11:05:43.876061 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-djbqg" event={"ID":"90315296-e6a2-4db3-aa3d-35af0d21a55e","Type":"ContainerStarted","Data":"d9d3a6f3db4e457acdf2eb32316c388679088a54bd3b8dee9f35a85b238f51c1"} Dec 05 11:05:43.876171 master-0 kubenswrapper[24928]: I1205 11:05:43.876132 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-djbqg" event={"ID":"90315296-e6a2-4db3-aa3d-35af0d21a55e","Type":"ContainerStarted","Data":"003031ae2b91b805735e82c853f5e2c92779ac6eca6d6175ce5ca5793a689b4f"} Dec 05 11:05:43.892182 master-0 kubenswrapper[24928]: I1205 11:05:43.892116 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64f947fff7-6npnx" event={"ID":"4bcd1d9a-8ff3-4577-8308-bd63c025d4e4","Type":"ContainerDied","Data":"bc3dc9f5b31d8d3a20a93174bc372f83eddfe539bb50d5c3be8a716b732ea96b"} Dec 05 11:05:43.892182 master-0 kubenswrapper[24928]: I1205 11:05:43.892187 24928 scope.go:117] "RemoveContainer" containerID="8642da1100670801b9eb83c9f0a447ca91642e8d3b1a211a1f3463a99d202cd1" Dec 05 11:05:43.892476 master-0 kubenswrapper[24928]: I1205 11:05:43.892326 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64f947fff7-6npnx" Dec 05 11:05:43.897645 master-0 kubenswrapper[24928]: I1205 11:05:43.897262 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:43.897645 master-0 kubenswrapper[24928]: I1205 11:05:43.897327 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-sb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:43.897645 master-0 kubenswrapper[24928]: I1205 11:05:43.897344 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:43.897645 master-0 kubenswrapper[24928]: I1205 11:05:43.897360 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plbnp\" (UniqueName: \"kubernetes.io/projected/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-kube-api-access-plbnp\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:43.897645 master-0 kubenswrapper[24928]: I1205 11:05:43.897376 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:43.900670 master-0 kubenswrapper[24928]: I1205 11:05:43.900594 24928 generic.go:334] "Generic (PLEG): container finished" podID="8cd6a2cb-720d-4653-bb1a-9b102506d7d3" containerID="fc890f12dae68037a73481d1b83307453d61eec457b30c9bacd0016b3e4ddc65" exitCode=0 Dec 05 11:05:43.900755 master-0 kubenswrapper[24928]: I1205 11:05:43.900708 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" event={"ID":"8cd6a2cb-720d-4653-bb1a-9b102506d7d3","Type":"ContainerDied","Data":"fc890f12dae68037a73481d1b83307453d61eec457b30c9bacd0016b3e4ddc65"} Dec 05 11:05:43.900800 master-0 kubenswrapper[24928]: I1205 11:05:43.900767 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" event={"ID":"8cd6a2cb-720d-4653-bb1a-9b102506d7d3","Type":"ContainerStarted","Data":"0b81c9b78b31c639fe1c47540edced7d4d0add5fdfb03726fc863e0ea2fe5ef0"} Dec 05 11:05:43.903499 master-0 kubenswrapper[24928]: I1205 11:05:43.903281 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "83a873c9-5adb-49e6-bb9c-cb9b205bff4c" (UID: "83a873c9-5adb-49e6-bb9c-cb9b205bff4c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:43.906479 master-0 kubenswrapper[24928]: I1205 11:05:43.906384 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-djbqg" podStartSLOduration=3.9063612340000002 podStartE2EDuration="3.906361234s" podCreationTimestamp="2025-12-05 11:05:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:05:43.899185725 +0000 UTC m=+1103.902379586" watchObservedRunningTime="2025-12-05 11:05:43.906361234 +0000 UTC m=+1103.909555085" Dec 05 11:05:43.908495 master-0 kubenswrapper[24928]: I1205 11:05:43.908400 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-config" (OuterVolumeSpecName: "config") pod "83a873c9-5adb-49e6-bb9c-cb9b205bff4c" (UID: "83a873c9-5adb-49e6-bb9c-cb9b205bff4c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:43.911707 master-0 kubenswrapper[24928]: I1205 11:05:43.911627 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x5nq4" event={"ID":"032f0939-20e7-46cb-9665-b1346b7e1b10","Type":"ContainerDied","Data":"234dc119535ca140563a43f2ddcb668024e8e834bb3944505096db4a7b86e233"} Dec 05 11:05:43.911835 master-0 kubenswrapper[24928]: I1205 11:05:43.911812 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x5nq4" Dec 05 11:05:43.946375 master-0 kubenswrapper[24928]: I1205 11:05:43.946331 24928 scope.go:117] "RemoveContainer" containerID="1ad8d4221e10a10b242cfb51bce3c73c08c96b11c2a121dce15f2c8ce277ce37" Dec 05 11:05:43.947030 master-0 kubenswrapper[24928]: I1205 11:05:43.946979 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-ac30-account-create-update-ngf65" event={"ID":"84942dea-64c3-403d-9d76-bbc2d9206f9f","Type":"ContainerStarted","Data":"cd6f3e3faf8ea4c6c8e88243dcd954d362933320b1173e42a2be96d848c1dff2"} Dec 05 11:05:43.976520 master-0 kubenswrapper[24928]: I1205 11:05:43.956087 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688fbc9bcc-96d2d"] Dec 05 11:05:43.976520 master-0 kubenswrapper[24928]: I1205 11:05:43.965343 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-rrvhc" event={"ID":"cd047ad1-2e39-47f6-ad8f-97eb080e7766","Type":"ContainerStarted","Data":"706172a6d379ad38ce7938c7ff6bac1f92546191b7fd0dc5fe68bae400a24cf4"} Dec 05 11:05:43.976520 master-0 kubenswrapper[24928]: W1205 11:05:43.971752 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod628a74ee_1afa_4e41_8b19_c6b0402a3368.slice/crio-9e8f4e99c9e0dc49a4f7df90a35a56440ae3111a49de2107d58d89aa2a9baad3 WatchSource:0}: Error finding container 9e8f4e99c9e0dc49a4f7df90a35a56440ae3111a49de2107d58d89aa2a9baad3: Status 404 returned error can't find the container with id 9e8f4e99c9e0dc49a4f7df90a35a56440ae3111a49de2107d58d89aa2a9baad3 Dec 05 11:05:43.976520 master-0 kubenswrapper[24928]: I1205 11:05:43.974106 24928 generic.go:334] "Generic (PLEG): container finished" podID="415910fe-1441-44d5-a220-322e3ba3e2e3" containerID="01008e247454c1f7b3a1ca02c81f438c58d10f34ab803c5c6a2b14645ce5b3c2" exitCode=0 Dec 05 11:05:43.976520 master-0 kubenswrapper[24928]: I1205 11:05:43.974177 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-x99hv" event={"ID":"415910fe-1441-44d5-a220-322e3ba3e2e3","Type":"ContainerDied","Data":"01008e247454c1f7b3a1ca02c81f438c58d10f34ab803c5c6a2b14645ce5b3c2"} Dec 05 11:05:43.977090 master-0 kubenswrapper[24928]: I1205 11:05:43.976996 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" Dec 05 11:05:43.978185 master-0 kubenswrapper[24928]: I1205 11:05:43.978110 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ff97bbb59-4fftj" event={"ID":"83a873c9-5adb-49e6-bb9c-cb9b205bff4c","Type":"ContainerDied","Data":"17c5fd5933838c3d550e293e2d67953432639362d09bfbb1d1cd5a32b0f60adf"} Dec 05 11:05:44.002880 master-0 kubenswrapper[24928]: I1205 11:05:44.002826 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:44.002880 master-0 kubenswrapper[24928]: I1205 11:05:44.002875 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/83a873c9-5adb-49e6-bb9c-cb9b205bff4c-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:44.039481 master-0 kubenswrapper[24928]: I1205 11:05:44.033808 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-ac30-account-create-update-ngf65" podStartSLOduration=4.033779002 podStartE2EDuration="4.033779002s" podCreationTimestamp="2025-12-05 11:05:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:05:43.980533804 +0000 UTC m=+1103.983727665" watchObservedRunningTime="2025-12-05 11:05:44.033779002 +0000 UTC m=+1104.036972863" Dec 05 11:05:44.087252 master-0 kubenswrapper[24928]: I1205 11:05:44.087186 24928 scope.go:117] "RemoveContainer" containerID="35606e6245a10ef961fcc399697eab348e9c687982934c3772679dc5994ccbb8" Dec 05 11:05:44.197927 master-0 kubenswrapper[24928]: I1205 11:05:44.197686 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64f947fff7-6npnx"] Dec 05 11:05:44.258445 master-0 kubenswrapper[24928]: I1205 11:05:44.257409 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64f947fff7-6npnx"] Dec 05 11:05:44.258445 master-0 kubenswrapper[24928]: I1205 11:05:44.257489 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x5nq4"] Dec 05 11:05:44.263864 master-0 kubenswrapper[24928]: I1205 11:05:44.263815 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x5nq4"] Dec 05 11:05:44.290312 master-0 kubenswrapper[24928]: I1205 11:05:44.290246 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:44.299269 master-0 kubenswrapper[24928]: I1205 11:05:44.299214 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:44.355688 master-0 kubenswrapper[24928]: I1205 11:05:44.348078 24928 scope.go:117] "RemoveContainer" containerID="199e2ba413ddeaeb33aee60c1f864ea1e644ea3137b0cc9fa2b3d04f4918b665" Dec 05 11:05:44.491936 master-0 kubenswrapper[24928]: I1205 11:05:44.491812 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ff97bbb59-4fftj"] Dec 05 11:05:44.556341 master-0 kubenswrapper[24928]: I1205 11:05:44.555483 24928 scope.go:117] "RemoveContainer" containerID="7032e3eb858119bfbbd19ea69283e2d6bc7ef3780f4127118ea99039c71dec81" Dec 05 11:05:44.572665 master-0 kubenswrapper[24928]: I1205 11:05:44.572565 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ff97bbb59-4fftj"] Dec 05 11:05:44.611636 master-0 kubenswrapper[24928]: I1205 11:05:44.606880 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:05:44.627539 master-0 kubenswrapper[24928]: I1205 11:05:44.623555 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:05:44.627539 master-0 kubenswrapper[24928]: E1205 11:05:44.624437 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[glance], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-ec941-default-internal-api-0" podUID="b2f49459-4e2d-4ada-b4af-4c73c515f1f3" Dec 05 11:05:44.663355 master-0 kubenswrapper[24928]: I1205 11:05:44.663274 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:44.738548 master-0 kubenswrapper[24928]: I1205 11:05:44.732356 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wz5q\" (UniqueName: \"kubernetes.io/projected/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-kube-api-access-2wz5q\") pod \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " Dec 05 11:05:44.738548 master-0 kubenswrapper[24928]: I1205 11:05:44.733296 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-config\") pod \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " Dec 05 11:05:44.738548 master-0 kubenswrapper[24928]: I1205 11:05:44.733605 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-sb\") pod \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " Dec 05 11:05:44.738548 master-0 kubenswrapper[24928]: I1205 11:05:44.733693 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-dns-svc\") pod \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " Dec 05 11:05:44.738548 master-0 kubenswrapper[24928]: I1205 11:05:44.733763 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-nb\") pod \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\" (UID: \"8cd6a2cb-720d-4653-bb1a-9b102506d7d3\") " Dec 05 11:05:44.758161 master-0 kubenswrapper[24928]: I1205 11:05:44.750741 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-kube-api-access-2wz5q" (OuterVolumeSpecName: "kube-api-access-2wz5q") pod "8cd6a2cb-720d-4653-bb1a-9b102506d7d3" (UID: "8cd6a2cb-720d-4653-bb1a-9b102506d7d3"). InnerVolumeSpecName "kube-api-access-2wz5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:44.770994 master-0 kubenswrapper[24928]: I1205 11:05:44.770945 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8cd6a2cb-720d-4653-bb1a-9b102506d7d3" (UID: "8cd6a2cb-720d-4653-bb1a-9b102506d7d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:44.778318 master-0 kubenswrapper[24928]: I1205 11:05:44.777655 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8cd6a2cb-720d-4653-bb1a-9b102506d7d3" (UID: "8cd6a2cb-720d-4653-bb1a-9b102506d7d3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:44.793711 master-0 kubenswrapper[24928]: I1205 11:05:44.793554 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-config" (OuterVolumeSpecName: "config") pod "8cd6a2cb-720d-4653-bb1a-9b102506d7d3" (UID: "8cd6a2cb-720d-4653-bb1a-9b102506d7d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:44.832615 master-0 kubenswrapper[24928]: I1205 11:05:44.831855 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8cd6a2cb-720d-4653-bb1a-9b102506d7d3" (UID: "8cd6a2cb-720d-4653-bb1a-9b102506d7d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:44.839081 master-0 kubenswrapper[24928]: I1205 11:05:44.837141 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:44.839081 master-0 kubenswrapper[24928]: I1205 11:05:44.837188 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wz5q\" (UniqueName: \"kubernetes.io/projected/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-kube-api-access-2wz5q\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:44.839081 master-0 kubenswrapper[24928]: I1205 11:05:44.837204 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:44.839081 master-0 kubenswrapper[24928]: I1205 11:05:44.837217 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-ovsdbserver-sb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:44.839081 master-0 kubenswrapper[24928]: I1205 11:05:44.837232 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8cd6a2cb-720d-4653-bb1a-9b102506d7d3-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:45.020907 master-0 kubenswrapper[24928]: I1205 11:05:45.020408 24928 generic.go:334] "Generic (PLEG): container finished" podID="628a74ee-1afa-4e41-8b19-c6b0402a3368" containerID="1076d61f988b2b7150a031c46ecbb7c89394dc2bfc88106697352ef40c99cbfe" exitCode=0 Dec 05 11:05:45.020907 master-0 kubenswrapper[24928]: I1205 11:05:45.020556 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" event={"ID":"628a74ee-1afa-4e41-8b19-c6b0402a3368","Type":"ContainerDied","Data":"1076d61f988b2b7150a031c46ecbb7c89394dc2bfc88106697352ef40c99cbfe"} Dec 05 11:05:45.020907 master-0 kubenswrapper[24928]: I1205 11:05:45.020595 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" event={"ID":"628a74ee-1afa-4e41-8b19-c6b0402a3368","Type":"ContainerStarted","Data":"9e8f4e99c9e0dc49a4f7df90a35a56440ae3111a49de2107d58d89aa2a9baad3"} Dec 05 11:05:45.028530 master-0 kubenswrapper[24928]: I1205 11:05:45.028471 24928 generic.go:334] "Generic (PLEG): container finished" podID="84942dea-64c3-403d-9d76-bbc2d9206f9f" containerID="cd6f3e3faf8ea4c6c8e88243dcd954d362933320b1173e42a2be96d848c1dff2" exitCode=0 Dec 05 11:05:45.028917 master-0 kubenswrapper[24928]: I1205 11:05:45.028561 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-ac30-account-create-update-ngf65" event={"ID":"84942dea-64c3-403d-9d76-bbc2d9206f9f","Type":"ContainerDied","Data":"cd6f3e3faf8ea4c6c8e88243dcd954d362933320b1173e42a2be96d848c1dff2"} Dec 05 11:05:45.040737 master-0 kubenswrapper[24928]: I1205 11:05:45.039015 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" Dec 05 11:05:45.040737 master-0 kubenswrapper[24928]: I1205 11:05:45.039142 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cbb8c9bdf-4rkl7" event={"ID":"8cd6a2cb-720d-4653-bb1a-9b102506d7d3","Type":"ContainerDied","Data":"0b81c9b78b31c639fe1c47540edced7d4d0add5fdfb03726fc863e0ea2fe5ef0"} Dec 05 11:05:45.040737 master-0 kubenswrapper[24928]: I1205 11:05:45.039244 24928 scope.go:117] "RemoveContainer" containerID="fc890f12dae68037a73481d1b83307453d61eec457b30c9bacd0016b3e4ddc65" Dec 05 11:05:45.040737 master-0 kubenswrapper[24928]: I1205 11:05:45.039457 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:45.299683 master-0 kubenswrapper[24928]: I1205 11:05:45.299379 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:05:45.317305 master-0 kubenswrapper[24928]: I1205 11:05:45.317047 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:45.379321 master-0 kubenswrapper[24928]: I1205 11:05:45.378297 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-logs\") pod \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " Dec 05 11:05:45.379321 master-0 kubenswrapper[24928]: I1205 11:05:45.378492 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-combined-ca-bundle\") pod \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " Dec 05 11:05:45.379321 master-0 kubenswrapper[24928]: I1205 11:05:45.378519 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-scripts\") pod \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " Dec 05 11:05:45.379321 master-0 kubenswrapper[24928]: I1205 11:05:45.378555 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7rb8r\" (UniqueName: \"kubernetes.io/projected/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-kube-api-access-7rb8r\") pod \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " Dec 05 11:05:45.379321 master-0 kubenswrapper[24928]: I1205 11:05:45.378674 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-config-data\") pod \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " Dec 05 11:05:45.379321 master-0 kubenswrapper[24928]: I1205 11:05:45.378752 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-httpd-run\") pod \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " Dec 05 11:05:45.379321 master-0 kubenswrapper[24928]: I1205 11:05:45.378769 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-logs" (OuterVolumeSpecName: "logs") pod "b2f49459-4e2d-4ada-b4af-4c73c515f1f3" (UID: "b2f49459-4e2d-4ada-b4af-4c73c515f1f3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:05:45.379654 master-0 kubenswrapper[24928]: I1205 11:05:45.379637 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:45.384738 master-0 kubenswrapper[24928]: I1205 11:05:45.380227 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b2f49459-4e2d-4ada-b4af-4c73c515f1f3" (UID: "b2f49459-4e2d-4ada-b4af-4c73c515f1f3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:05:45.404236 master-0 kubenswrapper[24928]: I1205 11:05:45.398223 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b2f49459-4e2d-4ada-b4af-4c73c515f1f3" (UID: "b2f49459-4e2d-4ada-b4af-4c73c515f1f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:45.412460 master-0 kubenswrapper[24928]: I1205 11:05:45.409401 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-scripts" (OuterVolumeSpecName: "scripts") pod "b2f49459-4e2d-4ada-b4af-4c73c515f1f3" (UID: "b2f49459-4e2d-4ada-b4af-4c73c515f1f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:45.434821 master-0 kubenswrapper[24928]: I1205 11:05:45.418372 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-config-data" (OuterVolumeSpecName: "config-data") pod "b2f49459-4e2d-4ada-b4af-4c73c515f1f3" (UID: "b2f49459-4e2d-4ada-b4af-4c73c515f1f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:45.434821 master-0 kubenswrapper[24928]: I1205 11:05:45.418856 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-kube-api-access-7rb8r" (OuterVolumeSpecName: "kube-api-access-7rb8r") pod "b2f49459-4e2d-4ada-b4af-4c73c515f1f3" (UID: "b2f49459-4e2d-4ada-b4af-4c73c515f1f3"). InnerVolumeSpecName "kube-api-access-7rb8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:45.457854 master-0 kubenswrapper[24928]: I1205 11:05:45.453283 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cbb8c9bdf-4rkl7"] Dec 05 11:05:45.472520 master-0 kubenswrapper[24928]: I1205 11:05:45.472452 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cbb8c9bdf-4rkl7"] Dec 05 11:05:45.487460 master-0 kubenswrapper[24928]: I1205 11:05:45.482461 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:45.487460 master-0 kubenswrapper[24928]: I1205 11:05:45.482500 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:45.487460 master-0 kubenswrapper[24928]: I1205 11:05:45.482513 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7rb8r\" (UniqueName: \"kubernetes.io/projected/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-kube-api-access-7rb8r\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:45.487460 master-0 kubenswrapper[24928]: I1205 11:05:45.482525 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:45.487460 master-0 kubenswrapper[24928]: I1205 11:05:45.482537 24928 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b2f49459-4e2d-4ada-b4af-4c73c515f1f3-httpd-run\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:45.657335 master-0 kubenswrapper[24928]: I1205 11:05:45.657286 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-x99hv" Dec 05 11:05:45.661147 master-0 kubenswrapper[24928]: I1205 11:05:45.661099 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:45.798312 master-0 kubenswrapper[24928]: I1205 11:05:45.798203 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4pf6\" (UniqueName: \"kubernetes.io/projected/415910fe-1441-44d5-a220-322e3ba3e2e3-kube-api-access-k4pf6\") pod \"415910fe-1441-44d5-a220-322e3ba3e2e3\" (UID: \"415910fe-1441-44d5-a220-322e3ba3e2e3\") " Dec 05 11:05:45.799047 master-0 kubenswrapper[24928]: I1205 11:05:45.798888 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/415910fe-1441-44d5-a220-322e3ba3e2e3-operator-scripts\") pod \"415910fe-1441-44d5-a220-322e3ba3e2e3\" (UID: \"415910fe-1441-44d5-a220-322e3ba3e2e3\") " Dec 05 11:05:45.799330 master-0 kubenswrapper[24928]: I1205 11:05:45.799202 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\" (UID: \"b2f49459-4e2d-4ada-b4af-4c73c515f1f3\") " Dec 05 11:05:45.799330 master-0 kubenswrapper[24928]: I1205 11:05:45.799219 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/415910fe-1441-44d5-a220-322e3ba3e2e3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "415910fe-1441-44d5-a220-322e3ba3e2e3" (UID: "415910fe-1441-44d5-a220-322e3ba3e2e3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:45.800266 master-0 kubenswrapper[24928]: I1205 11:05:45.800220 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/415910fe-1441-44d5-a220-322e3ba3e2e3-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:45.803237 master-0 kubenswrapper[24928]: I1205 11:05:45.803191 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/415910fe-1441-44d5-a220-322e3ba3e2e3-kube-api-access-k4pf6" (OuterVolumeSpecName: "kube-api-access-k4pf6") pod "415910fe-1441-44d5-a220-322e3ba3e2e3" (UID: "415910fe-1441-44d5-a220-322e3ba3e2e3"). InnerVolumeSpecName "kube-api-access-k4pf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:45.819538 master-0 kubenswrapper[24928]: I1205 11:05:45.819462 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880" (OuterVolumeSpecName: "glance") pod "b2f49459-4e2d-4ada-b4af-4c73c515f1f3" (UID: "b2f49459-4e2d-4ada-b4af-4c73c515f1f3"). InnerVolumeSpecName "pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 11:05:45.902541 master-0 kubenswrapper[24928]: I1205 11:05:45.902414 24928 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") on node \"master-0\" " Dec 05 11:05:45.902541 master-0 kubenswrapper[24928]: I1205 11:05:45.902476 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4pf6\" (UniqueName: \"kubernetes.io/projected/415910fe-1441-44d5-a220-322e3ba3e2e3-kube-api-access-k4pf6\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:45.968823 master-0 kubenswrapper[24928]: I1205 11:05:45.968069 24928 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 11:05:45.968823 master-0 kubenswrapper[24928]: I1205 11:05:45.968262 24928 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01" (UniqueName: "kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880") on node "master-0" Dec 05 11:05:46.005208 master-0 kubenswrapper[24928]: I1205 11:05:46.005148 24928 reconciler_common.go:293] "Volume detached for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:46.067950 master-0 kubenswrapper[24928]: I1205 11:05:46.067851 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" event={"ID":"628a74ee-1afa-4e41-8b19-c6b0402a3368","Type":"ContainerStarted","Data":"69a82a43ffa8d33e5be2b5b8b95406f7cecc72aa4a8fec74bdea08272875fac1"} Dec 05 11:05:46.067950 master-0 kubenswrapper[24928]: I1205 11:05:46.067952 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:46.073977 master-0 kubenswrapper[24928]: I1205 11:05:46.073698 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-x99hv" event={"ID":"415910fe-1441-44d5-a220-322e3ba3e2e3","Type":"ContainerDied","Data":"021e7f11b0e27fa74a4cb3e7e4e907a609e13912b8ae022d6031e3fc609c6a9f"} Dec 05 11:05:46.073977 master-0 kubenswrapper[24928]: I1205 11:05:46.073746 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="021e7f11b0e27fa74a4cb3e7e4e907a609e13912b8ae022d6031e3fc609c6a9f" Dec 05 11:05:46.073977 master-0 kubenswrapper[24928]: I1205 11:05:46.073835 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-x99hv" Dec 05 11:05:46.075905 master-0 kubenswrapper[24928]: I1205 11:05:46.075675 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"577d038e-b40d-43e5-a6ad-ad4e278a3213","Type":"ContainerStarted","Data":"fbeb038a6a5ced2e15023521c841c04a9b7c72a59126863e427829e092f86127"} Dec 05 11:05:46.075905 master-0 kubenswrapper[24928]: I1205 11:05:46.075784 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.116776 master-0 kubenswrapper[24928]: I1205 11:05:46.116654 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" podStartSLOduration=4.116627382 podStartE2EDuration="4.116627382s" podCreationTimestamp="2025-12-05 11:05:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:05:46.08969612 +0000 UTC m=+1106.092889981" watchObservedRunningTime="2025-12-05 11:05:46.116627382 +0000 UTC m=+1106.119821243" Dec 05 11:05:46.190624 master-0 kubenswrapper[24928]: I1205 11:05:46.188982 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:05:46.242088 master-0 kubenswrapper[24928]: I1205 11:05:46.231068 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="032f0939-20e7-46cb-9665-b1346b7e1b10" path="/var/lib/kubelet/pods/032f0939-20e7-46cb-9665-b1346b7e1b10/volumes" Dec 05 11:05:46.242088 master-0 kubenswrapper[24928]: I1205 11:05:46.232341 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bcd1d9a-8ff3-4577-8308-bd63c025d4e4" path="/var/lib/kubelet/pods/4bcd1d9a-8ff3-4577-8308-bd63c025d4e4/volumes" Dec 05 11:05:46.242088 master-0 kubenswrapper[24928]: I1205 11:05:46.234003 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a873c9-5adb-49e6-bb9c-cb9b205bff4c" path="/var/lib/kubelet/pods/83a873c9-5adb-49e6-bb9c-cb9b205bff4c/volumes" Dec 05 11:05:46.242088 master-0 kubenswrapper[24928]: I1205 11:05:46.236757 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cd6a2cb-720d-4653-bb1a-9b102506d7d3" path="/var/lib/kubelet/pods/8cd6a2cb-720d-4653-bb1a-9b102506d7d3/volumes" Dec 05 11:05:46.242088 master-0 kubenswrapper[24928]: I1205 11:05:46.237350 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: I1205 11:05:46.246027 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: E1205 11:05:46.248121 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a873c9-5adb-49e6-bb9c-cb9b205bff4c" containerName="init" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: I1205 11:05:46.248159 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a873c9-5adb-49e6-bb9c-cb9b205bff4c" containerName="init" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: E1205 11:05:46.248188 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bcd1d9a-8ff3-4577-8308-bd63c025d4e4" containerName="init" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: I1205 11:05:46.248195 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bcd1d9a-8ff3-4577-8308-bd63c025d4e4" containerName="init" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: E1205 11:05:46.248238 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415910fe-1441-44d5-a220-322e3ba3e2e3" containerName="mariadb-database-create" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: I1205 11:05:46.248244 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="415910fe-1441-44d5-a220-322e3ba3e2e3" containerName="mariadb-database-create" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: E1205 11:05:46.248264 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cd6a2cb-720d-4653-bb1a-9b102506d7d3" containerName="init" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: I1205 11:05:46.248270 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cd6a2cb-720d-4653-bb1a-9b102506d7d3" containerName="init" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: I1205 11:05:46.248737 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="415910fe-1441-44d5-a220-322e3ba3e2e3" containerName="mariadb-database-create" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: I1205 11:05:46.248764 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bcd1d9a-8ff3-4577-8308-bd63c025d4e4" containerName="init" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: I1205 11:05:46.248774 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a873c9-5adb-49e6-bb9c-cb9b205bff4c" containerName="init" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: I1205 11:05:46.248795 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cd6a2cb-720d-4653-bb1a-9b102506d7d3" containerName="init" Dec 05 11:05:46.252414 master-0 kubenswrapper[24928]: I1205 11:05:46.250834 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.257835 master-0 kubenswrapper[24928]: I1205 11:05:46.257171 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-ec941-default-internal-config-data" Dec 05 11:05:46.311505 master-0 kubenswrapper[24928]: I1205 11:05:46.308280 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:05:46.322683 master-0 kubenswrapper[24928]: I1205 11:05:46.319480 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.322683 master-0 kubenswrapper[24928]: I1205 11:05:46.319599 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.322683 master-0 kubenswrapper[24928]: I1205 11:05:46.319633 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.322683 master-0 kubenswrapper[24928]: I1205 11:05:46.320032 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w8c7\" (UniqueName: \"kubernetes.io/projected/c257c518-4db1-4ecd-abac-6bd4578c77d3-kube-api-access-9w8c7\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.322683 master-0 kubenswrapper[24928]: I1205 11:05:46.320278 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.322683 master-0 kubenswrapper[24928]: I1205 11:05:46.320320 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.322683 master-0 kubenswrapper[24928]: I1205 11:05:46.320596 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.422858 master-0 kubenswrapper[24928]: I1205 11:05:46.422694 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w8c7\" (UniqueName: \"kubernetes.io/projected/c257c518-4db1-4ecd-abac-6bd4578c77d3-kube-api-access-9w8c7\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.422858 master-0 kubenswrapper[24928]: I1205 11:05:46.422825 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.422858 master-0 kubenswrapper[24928]: I1205 11:05:46.422852 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.423133 master-0 kubenswrapper[24928]: I1205 11:05:46.422896 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.423133 master-0 kubenswrapper[24928]: I1205 11:05:46.423002 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.423133 master-0 kubenswrapper[24928]: I1205 11:05:46.423051 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.423133 master-0 kubenswrapper[24928]: I1205 11:05:46.423084 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.423757 master-0 kubenswrapper[24928]: I1205 11:05:46.423708 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.427138 master-0 kubenswrapper[24928]: I1205 11:05:46.426939 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.432035 master-0 kubenswrapper[24928]: I1205 11:05:46.431955 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.433117 master-0 kubenswrapper[24928]: I1205 11:05:46.433087 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:05:46.433206 master-0 kubenswrapper[24928]: I1205 11:05:46.433139 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/f829e9071eee306944ccf41eabbcf74a1473d7031ba724c821f61b9a30a289e9/globalmount\"" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.435523 master-0 kubenswrapper[24928]: I1205 11:05:46.434251 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.435523 master-0 kubenswrapper[24928]: I1205 11:05:46.435036 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:46.446175 master-0 kubenswrapper[24928]: I1205 11:05:46.445602 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w8c7\" (UniqueName: \"kubernetes.io/projected/c257c518-4db1-4ecd-abac-6bd4578c77d3-kube-api-access-9w8c7\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:47.092445 master-0 kubenswrapper[24928]: I1205 11:05:47.091749 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"577d038e-b40d-43e5-a6ad-ad4e278a3213","Type":"ContainerStarted","Data":"b64446ffcb31f96aff3c25293f9888d8c03de5e18dd61a127765110d92a94e5d"} Dec 05 11:05:47.092445 master-0 kubenswrapper[24928]: I1205 11:05:47.091809 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"577d038e-b40d-43e5-a6ad-ad4e278a3213","Type":"ContainerStarted","Data":"c1a2d44c8f4b865d29401fec7830bdc9df6ee3ba9d18ca63cdce0eca1256b656"} Dec 05 11:05:47.092445 master-0 kubenswrapper[24928]: I1205 11:05:47.092050 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-ec941-default-external-api-0" podUID="577d038e-b40d-43e5-a6ad-ad4e278a3213" containerName="glance-log" containerID="cri-o://c1a2d44c8f4b865d29401fec7830bdc9df6ee3ba9d18ca63cdce0eca1256b656" gracePeriod=30 Dec 05 11:05:47.092445 master-0 kubenswrapper[24928]: I1205 11:05:47.092117 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-ec941-default-external-api-0" podUID="577d038e-b40d-43e5-a6ad-ad4e278a3213" containerName="glance-httpd" containerID="cri-o://b64446ffcb31f96aff3c25293f9888d8c03de5e18dd61a127765110d92a94e5d" gracePeriod=30 Dec 05 11:05:47.115057 master-0 kubenswrapper[24928]: I1205 11:05:47.113876 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-ec941-default-external-api-0" podStartSLOduration=7.113854284 podStartE2EDuration="7.113854284s" podCreationTimestamp="2025-12-05 11:05:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:05:47.112936491 +0000 UTC m=+1107.116130342" watchObservedRunningTime="2025-12-05 11:05:47.113854284 +0000 UTC m=+1107.117048135" Dec 05 11:05:47.561367 master-0 kubenswrapper[24928]: I1205 11:05:47.561304 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:47.565636 master-0 kubenswrapper[24928]: I1205 11:05:47.565572 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:05:48.124006 master-0 kubenswrapper[24928]: I1205 11:05:48.123923 24928 generic.go:334] "Generic (PLEG): container finished" podID="6140cfa3-11aa-4550-a8dc-ed81dcd087be" containerID="147d4e2b823107dd73c2695e1569f383bc3f459b9c676b6a321be91f112e4b3e" exitCode=0 Dec 05 11:05:48.124006 master-0 kubenswrapper[24928]: I1205 11:05:48.123969 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r7lkl" event={"ID":"6140cfa3-11aa-4550-a8dc-ed81dcd087be","Type":"ContainerDied","Data":"147d4e2b823107dd73c2695e1569f383bc3f459b9c676b6a321be91f112e4b3e"} Dec 05 11:05:48.128796 master-0 kubenswrapper[24928]: I1205 11:05:48.128755 24928 generic.go:334] "Generic (PLEG): container finished" podID="577d038e-b40d-43e5-a6ad-ad4e278a3213" containerID="b64446ffcb31f96aff3c25293f9888d8c03de5e18dd61a127765110d92a94e5d" exitCode=143 Dec 05 11:05:48.128796 master-0 kubenswrapper[24928]: I1205 11:05:48.128789 24928 generic.go:334] "Generic (PLEG): container finished" podID="577d038e-b40d-43e5-a6ad-ad4e278a3213" containerID="c1a2d44c8f4b865d29401fec7830bdc9df6ee3ba9d18ca63cdce0eca1256b656" exitCode=143 Dec 05 11:05:48.129000 master-0 kubenswrapper[24928]: I1205 11:05:48.128809 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"577d038e-b40d-43e5-a6ad-ad4e278a3213","Type":"ContainerDied","Data":"b64446ffcb31f96aff3c25293f9888d8c03de5e18dd61a127765110d92a94e5d"} Dec 05 11:05:48.129000 master-0 kubenswrapper[24928]: I1205 11:05:48.128833 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"577d038e-b40d-43e5-a6ad-ad4e278a3213","Type":"ContainerDied","Data":"c1a2d44c8f4b865d29401fec7830bdc9df6ee3ba9d18ca63cdce0eca1256b656"} Dec 05 11:05:48.220460 master-0 kubenswrapper[24928]: I1205 11:05:48.220206 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2f49459-4e2d-4ada-b4af-4c73c515f1f3" path="/var/lib/kubelet/pods/b2f49459-4e2d-4ada-b4af-4c73c515f1f3/volumes" Dec 05 11:05:48.901951 master-0 kubenswrapper[24928]: I1205 11:05:48.901840 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-ac30-account-create-update-ngf65" Dec 05 11:05:49.017497 master-0 kubenswrapper[24928]: I1205 11:05:49.016084 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84942dea-64c3-403d-9d76-bbc2d9206f9f-operator-scripts\") pod \"84942dea-64c3-403d-9d76-bbc2d9206f9f\" (UID: \"84942dea-64c3-403d-9d76-bbc2d9206f9f\") " Dec 05 11:05:49.017497 master-0 kubenswrapper[24928]: I1205 11:05:49.016156 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pslsc\" (UniqueName: \"kubernetes.io/projected/84942dea-64c3-403d-9d76-bbc2d9206f9f-kube-api-access-pslsc\") pod \"84942dea-64c3-403d-9d76-bbc2d9206f9f\" (UID: \"84942dea-64c3-403d-9d76-bbc2d9206f9f\") " Dec 05 11:05:49.019517 master-0 kubenswrapper[24928]: I1205 11:05:49.018565 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/84942dea-64c3-403d-9d76-bbc2d9206f9f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "84942dea-64c3-403d-9d76-bbc2d9206f9f" (UID: "84942dea-64c3-403d-9d76-bbc2d9206f9f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:05:49.054446 master-0 kubenswrapper[24928]: I1205 11:05:49.046193 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84942dea-64c3-403d-9d76-bbc2d9206f9f-kube-api-access-pslsc" (OuterVolumeSpecName: "kube-api-access-pslsc") pod "84942dea-64c3-403d-9d76-bbc2d9206f9f" (UID: "84942dea-64c3-403d-9d76-bbc2d9206f9f"). InnerVolumeSpecName "kube-api-access-pslsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:49.118412 master-0 kubenswrapper[24928]: I1205 11:05:49.118299 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:05:49.126196 master-0 kubenswrapper[24928]: I1205 11:05:49.126122 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/84942dea-64c3-403d-9d76-bbc2d9206f9f-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:49.126842 master-0 kubenswrapper[24928]: I1205 11:05:49.126816 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pslsc\" (UniqueName: \"kubernetes.io/projected/84942dea-64c3-403d-9d76-bbc2d9206f9f-kube-api-access-pslsc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:49.209379 master-0 kubenswrapper[24928]: I1205 11:05:49.208619 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-ac30-account-create-update-ngf65" event={"ID":"84942dea-64c3-403d-9d76-bbc2d9206f9f","Type":"ContainerDied","Data":"949177e2d5f42a0e3bbcecac140fa1c4deadacef7f1617d32fe60dcb86112cf6"} Dec 05 11:05:49.209379 master-0 kubenswrapper[24928]: I1205 11:05:49.208688 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="949177e2d5f42a0e3bbcecac140fa1c4deadacef7f1617d32fe60dcb86112cf6" Dec 05 11:05:49.209379 master-0 kubenswrapper[24928]: I1205 11:05:49.208777 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-ac30-account-create-update-ngf65" Dec 05 11:05:49.234950 master-0 kubenswrapper[24928]: I1205 11:05:49.227797 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-rrvhc" event={"ID":"cd047ad1-2e39-47f6-ad8f-97eb080e7766","Type":"ContainerStarted","Data":"0fd5eac4fad287c65188ecc90c8530f87659bfbe84a9ec67de0571e498c18967"} Dec 05 11:05:49.268837 master-0 kubenswrapper[24928]: I1205 11:05:49.268314 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-rrvhc" podStartSLOduration=3.590715652 podStartE2EDuration="9.268293779s" podCreationTimestamp="2025-12-05 11:05:40 +0000 UTC" firstStartedPulling="2025-12-05 11:05:42.978570203 +0000 UTC m=+1102.981764044" lastFinishedPulling="2025-12-05 11:05:48.65614832 +0000 UTC m=+1108.659342171" observedRunningTime="2025-12-05 11:05:49.264051393 +0000 UTC m=+1109.267245254" watchObservedRunningTime="2025-12-05 11:05:49.268293779 +0000 UTC m=+1109.271487640" Dec 05 11:05:49.299654 master-0 kubenswrapper[24928]: I1205 11:05:49.299598 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:49.300873 master-0 kubenswrapper[24928]: I1205 11:05:49.300829 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:05:49.432730 master-0 kubenswrapper[24928]: I1205 11:05:49.432668 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-logs\") pod \"577d038e-b40d-43e5-a6ad-ad4e278a3213\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " Dec 05 11:05:49.432939 master-0 kubenswrapper[24928]: I1205 11:05:49.432767 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-config-data\") pod \"577d038e-b40d-43e5-a6ad-ad4e278a3213\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " Dec 05 11:05:49.432939 master-0 kubenswrapper[24928]: I1205 11:05:49.432844 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcm2z\" (UniqueName: \"kubernetes.io/projected/577d038e-b40d-43e5-a6ad-ad4e278a3213-kube-api-access-vcm2z\") pod \"577d038e-b40d-43e5-a6ad-ad4e278a3213\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " Dec 05 11:05:49.433014 master-0 kubenswrapper[24928]: I1205 11:05:49.432966 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-httpd-run\") pod \"577d038e-b40d-43e5-a6ad-ad4e278a3213\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " Dec 05 11:05:49.433067 master-0 kubenswrapper[24928]: I1205 11:05:49.433044 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-combined-ca-bundle\") pod \"577d038e-b40d-43e5-a6ad-ad4e278a3213\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " Dec 05 11:05:49.433106 master-0 kubenswrapper[24928]: I1205 11:05:49.433070 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-scripts\") pod \"577d038e-b40d-43e5-a6ad-ad4e278a3213\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " Dec 05 11:05:49.433440 master-0 kubenswrapper[24928]: I1205 11:05:49.433157 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"577d038e-b40d-43e5-a6ad-ad4e278a3213\" (UID: \"577d038e-b40d-43e5-a6ad-ad4e278a3213\") " Dec 05 11:05:49.433440 master-0 kubenswrapper[24928]: I1205 11:05:49.433307 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-logs" (OuterVolumeSpecName: "logs") pod "577d038e-b40d-43e5-a6ad-ad4e278a3213" (UID: "577d038e-b40d-43e5-a6ad-ad4e278a3213"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:05:49.434530 master-0 kubenswrapper[24928]: I1205 11:05:49.433585 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "577d038e-b40d-43e5-a6ad-ad4e278a3213" (UID: "577d038e-b40d-43e5-a6ad-ad4e278a3213"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:05:49.434530 master-0 kubenswrapper[24928]: I1205 11:05:49.434068 24928 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-httpd-run\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:49.434530 master-0 kubenswrapper[24928]: I1205 11:05:49.434086 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/577d038e-b40d-43e5-a6ad-ad4e278a3213-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:49.440390 master-0 kubenswrapper[24928]: I1205 11:05:49.439718 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-scripts" (OuterVolumeSpecName: "scripts") pod "577d038e-b40d-43e5-a6ad-ad4e278a3213" (UID: "577d038e-b40d-43e5-a6ad-ad4e278a3213"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:49.440390 master-0 kubenswrapper[24928]: I1205 11:05:49.439767 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/577d038e-b40d-43e5-a6ad-ad4e278a3213-kube-api-access-vcm2z" (OuterVolumeSpecName: "kube-api-access-vcm2z") pod "577d038e-b40d-43e5-a6ad-ad4e278a3213" (UID: "577d038e-b40d-43e5-a6ad-ad4e278a3213"). InnerVolumeSpecName "kube-api-access-vcm2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:49.471695 master-0 kubenswrapper[24928]: I1205 11:05:49.471586 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19" (OuterVolumeSpecName: "glance") pod "577d038e-b40d-43e5-a6ad-ad4e278a3213" (UID: "577d038e-b40d-43e5-a6ad-ad4e278a3213"). InnerVolumeSpecName "pvc-2a5fd065-f556-4411-b867-061babb4188e". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 11:05:49.482956 master-0 kubenswrapper[24928]: I1205 11:05:49.482886 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "577d038e-b40d-43e5-a6ad-ad4e278a3213" (UID: "577d038e-b40d-43e5-a6ad-ad4e278a3213"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:49.515203 master-0 kubenswrapper[24928]: I1205 11:05:49.515026 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-config-data" (OuterVolumeSpecName: "config-data") pod "577d038e-b40d-43e5-a6ad-ad4e278a3213" (UID: "577d038e-b40d-43e5-a6ad-ad4e278a3213"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:49.538492 master-0 kubenswrapper[24928]: I1205 11:05:49.536716 24928 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") on node \"master-0\" " Dec 05 11:05:49.538492 master-0 kubenswrapper[24928]: I1205 11:05:49.536778 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:49.538492 master-0 kubenswrapper[24928]: I1205 11:05:49.536793 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:49.538492 master-0 kubenswrapper[24928]: I1205 11:05:49.536807 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/577d038e-b40d-43e5-a6ad-ad4e278a3213-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:49.538492 master-0 kubenswrapper[24928]: I1205 11:05:49.536821 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcm2z\" (UniqueName: \"kubernetes.io/projected/577d038e-b40d-43e5-a6ad-ad4e278a3213-kube-api-access-vcm2z\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:49.577914 master-0 kubenswrapper[24928]: I1205 11:05:49.575767 24928 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 11:05:49.577914 master-0 kubenswrapper[24928]: I1205 11:05:49.576032 24928 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2a5fd065-f556-4411-b867-061babb4188e" (UniqueName: "kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19") on node "master-0" Dec 05 11:05:49.640309 master-0 kubenswrapper[24928]: I1205 11:05:49.640101 24928 reconciler_common.go:293] "Volume detached for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:49.879549 master-0 kubenswrapper[24928]: I1205 11:05:49.879416 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:49.963436 master-0 kubenswrapper[24928]: I1205 11:05:49.962344 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-config-data\") pod \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " Dec 05 11:05:49.963436 master-0 kubenswrapper[24928]: I1205 11:05:49.962792 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-combined-ca-bundle\") pod \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " Dec 05 11:05:49.963436 master-0 kubenswrapper[24928]: I1205 11:05:49.962905 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-fernet-keys\") pod \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " Dec 05 11:05:49.963436 master-0 kubenswrapper[24928]: I1205 11:05:49.963016 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-credential-keys\") pod \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " Dec 05 11:05:49.963436 master-0 kubenswrapper[24928]: I1205 11:05:49.963173 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-scripts\") pod \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " Dec 05 11:05:49.970949 master-0 kubenswrapper[24928]: I1205 11:05:49.963471 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drztt\" (UniqueName: \"kubernetes.io/projected/6140cfa3-11aa-4550-a8dc-ed81dcd087be-kube-api-access-drztt\") pod \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\" (UID: \"6140cfa3-11aa-4550-a8dc-ed81dcd087be\") " Dec 05 11:05:49.970949 master-0 kubenswrapper[24928]: I1205 11:05:49.968298 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-scripts" (OuterVolumeSpecName: "scripts") pod "6140cfa3-11aa-4550-a8dc-ed81dcd087be" (UID: "6140cfa3-11aa-4550-a8dc-ed81dcd087be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:49.970949 master-0 kubenswrapper[24928]: I1205 11:05:49.970607 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6140cfa3-11aa-4550-a8dc-ed81dcd087be" (UID: "6140cfa3-11aa-4550-a8dc-ed81dcd087be"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:49.991504 master-0 kubenswrapper[24928]: I1205 11:05:49.991394 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6140cfa3-11aa-4550-a8dc-ed81dcd087be" (UID: "6140cfa3-11aa-4550-a8dc-ed81dcd087be"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:49.996075 master-0 kubenswrapper[24928]: I1205 11:05:49.995795 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6140cfa3-11aa-4550-a8dc-ed81dcd087be-kube-api-access-drztt" (OuterVolumeSpecName: "kube-api-access-drztt") pod "6140cfa3-11aa-4550-a8dc-ed81dcd087be" (UID: "6140cfa3-11aa-4550-a8dc-ed81dcd087be"). InnerVolumeSpecName "kube-api-access-drztt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:05:50.004653 master-0 kubenswrapper[24928]: I1205 11:05:50.004452 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:50.033109 master-0 kubenswrapper[24928]: I1205 11:05:50.033016 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6140cfa3-11aa-4550-a8dc-ed81dcd087be" (UID: "6140cfa3-11aa-4550-a8dc-ed81dcd087be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:50.074536 master-0 kubenswrapper[24928]: I1205 11:05:50.068024 24928 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-credential-keys\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:50.074536 master-0 kubenswrapper[24928]: I1205 11:05:50.068105 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:50.074536 master-0 kubenswrapper[24928]: I1205 11:05:50.068119 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drztt\" (UniqueName: \"kubernetes.io/projected/6140cfa3-11aa-4550-a8dc-ed81dcd087be-kube-api-access-drztt\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:50.074536 master-0 kubenswrapper[24928]: I1205 11:05:50.068302 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:50.074536 master-0 kubenswrapper[24928]: I1205 11:05:50.068315 24928 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-fernet-keys\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:50.078564 master-0 kubenswrapper[24928]: I1205 11:05:50.076921 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:05:50.089233 master-0 kubenswrapper[24928]: I1205 11:05:50.089169 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-config-data" (OuterVolumeSpecName: "config-data") pod "6140cfa3-11aa-4550-a8dc-ed81dcd087be" (UID: "6140cfa3-11aa-4550-a8dc-ed81dcd087be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:05:50.171494 master-0 kubenswrapper[24928]: I1205 11:05:50.171132 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6140cfa3-11aa-4550-a8dc-ed81dcd087be-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:05:50.263307 master-0 kubenswrapper[24928]: I1205 11:05:50.263218 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r7lkl" event={"ID":"6140cfa3-11aa-4550-a8dc-ed81dcd087be","Type":"ContainerDied","Data":"95e5cfae88fd5e3131dff9262f4a9817a8cc318f85370c21381382497a509fbd"} Dec 05 11:05:50.263307 master-0 kubenswrapper[24928]: I1205 11:05:50.263301 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95e5cfae88fd5e3131dff9262f4a9817a8cc318f85370c21381382497a509fbd" Dec 05 11:05:50.263499 master-0 kubenswrapper[24928]: I1205 11:05:50.263259 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r7lkl" Dec 05 11:05:50.268439 master-0 kubenswrapper[24928]: I1205 11:05:50.266889 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"577d038e-b40d-43e5-a6ad-ad4e278a3213","Type":"ContainerDied","Data":"fbeb038a6a5ced2e15023521c841c04a9b7c72a59126863e427829e092f86127"} Dec 05 11:05:50.268439 master-0 kubenswrapper[24928]: I1205 11:05:50.266968 24928 scope.go:117] "RemoveContainer" containerID="b64446ffcb31f96aff3c25293f9888d8c03de5e18dd61a127765110d92a94e5d" Dec 05 11:05:50.268439 master-0 kubenswrapper[24928]: I1205 11:05:50.267175 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.276957 master-0 kubenswrapper[24928]: I1205 11:05:50.276908 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"c257c518-4db1-4ecd-abac-6bd4578c77d3","Type":"ContainerStarted","Data":"acab22f705229821317c8e656f6a30e41c6b91f42ae6d7d71598850b764f4626"} Dec 05 11:05:50.277097 master-0 kubenswrapper[24928]: I1205 11:05:50.276965 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"c257c518-4db1-4ecd-abac-6bd4578c77d3","Type":"ContainerStarted","Data":"845c2dd596934a91b79bd108e6d04ccfc513681cf6a760708c5de4c3208f38bd"} Dec 05 11:05:50.284901 master-0 kubenswrapper[24928]: I1205 11:05:50.284833 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vdmr2"] Dec 05 11:05:50.303564 master-0 kubenswrapper[24928]: I1205 11:05:50.303452 24928 scope.go:117] "RemoveContainer" containerID="c1a2d44c8f4b865d29401fec7830bdc9df6ee3ba9d18ca63cdce0eca1256b656" Dec 05 11:05:50.326564 master-0 kubenswrapper[24928]: I1205 11:05:50.326490 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:05:50.334476 master-0 kubenswrapper[24928]: I1205 11:05:50.334347 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-r7lkl"] Dec 05 11:05:50.350496 master-0 kubenswrapper[24928]: I1205 11:05:50.350376 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-r7lkl"] Dec 05 11:05:50.364212 master-0 kubenswrapper[24928]: I1205 11:05:50.363533 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.376949 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: E1205 11:05:50.380983 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="577d038e-b40d-43e5-a6ad-ad4e278a3213" containerName="glance-httpd" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.381027 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="577d038e-b40d-43e5-a6ad-ad4e278a3213" containerName="glance-httpd" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: E1205 11:05:50.381040 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6140cfa3-11aa-4550-a8dc-ed81dcd087be" containerName="keystone-bootstrap" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.381049 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6140cfa3-11aa-4550-a8dc-ed81dcd087be" containerName="keystone-bootstrap" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: E1205 11:05:50.381102 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84942dea-64c3-403d-9d76-bbc2d9206f9f" containerName="mariadb-account-create-update" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.381112 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="84942dea-64c3-403d-9d76-bbc2d9206f9f" containerName="mariadb-account-create-update" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: E1205 11:05:50.381121 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="577d038e-b40d-43e5-a6ad-ad4e278a3213" containerName="glance-log" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.381128 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="577d038e-b40d-43e5-a6ad-ad4e278a3213" containerName="glance-log" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.381366 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="6140cfa3-11aa-4550-a8dc-ed81dcd087be" containerName="keystone-bootstrap" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.381436 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="84942dea-64c3-403d-9d76-bbc2d9206f9f" containerName="mariadb-account-create-update" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.381455 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="577d038e-b40d-43e5-a6ad-ad4e278a3213" containerName="glance-log" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.381481 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="577d038e-b40d-43e5-a6ad-ad4e278a3213" containerName="glance-httpd" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.382611 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.385125 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 11:05:50.389492 master-0 kubenswrapper[24928]: I1205 11:05:50.385144 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-ec941-default-external-config-data" Dec 05 11:05:50.414269 master-0 kubenswrapper[24928]: I1205 11:05:50.414200 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:05:50.440311 master-0 kubenswrapper[24928]: I1205 11:05:50.428055 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2wx5d"] Dec 05 11:05:50.440311 master-0 kubenswrapper[24928]: I1205 11:05:50.429676 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.440311 master-0 kubenswrapper[24928]: I1205 11:05:50.434468 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 11:05:50.440311 master-0 kubenswrapper[24928]: I1205 11:05:50.434713 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 11:05:50.440311 master-0 kubenswrapper[24928]: I1205 11:05:50.434879 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 11:05:50.440311 master-0 kubenswrapper[24928]: I1205 11:05:50.435617 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 11:05:50.444000 master-0 kubenswrapper[24928]: I1205 11:05:50.443923 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2wx5d"] Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.484184 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ktrt\" (UniqueName: \"kubernetes.io/projected/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-kube-api-access-6ktrt\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.484266 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-config-data\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.484319 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-config-data\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.484380 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw8sl\" (UniqueName: \"kubernetes.io/projected/11a17674-7ccb-402d-9910-7ec8797f38b8-kube-api-access-vw8sl\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.484481 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-fernet-keys\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.484544 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-combined-ca-bundle\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.485105 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-credential-keys\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.485300 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.485358 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-scripts\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.485576 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-httpd-run\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.485640 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-public-tls-certs\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.485672 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-combined-ca-bundle\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.485711 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-logs\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.486766 master-0 kubenswrapper[24928]: I1205 11:05:50.485734 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-scripts\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.589127 master-0 kubenswrapper[24928]: I1205 11:05:50.588722 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-config-data\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.589127 master-0 kubenswrapper[24928]: I1205 11:05:50.588811 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw8sl\" (UniqueName: \"kubernetes.io/projected/11a17674-7ccb-402d-9910-7ec8797f38b8-kube-api-access-vw8sl\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.589127 master-0 kubenswrapper[24928]: I1205 11:05:50.588843 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-fernet-keys\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.589127 master-0 kubenswrapper[24928]: I1205 11:05:50.588878 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-combined-ca-bundle\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.589127 master-0 kubenswrapper[24928]: I1205 11:05:50.588922 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-credential-keys\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.589127 master-0 kubenswrapper[24928]: I1205 11:05:50.588962 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.589127 master-0 kubenswrapper[24928]: I1205 11:05:50.588997 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-scripts\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.589127 master-0 kubenswrapper[24928]: I1205 11:05:50.589101 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-httpd-run\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.589127 master-0 kubenswrapper[24928]: I1205 11:05:50.589140 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-public-tls-certs\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.590467 master-0 kubenswrapper[24928]: I1205 11:05:50.589163 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-combined-ca-bundle\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.590467 master-0 kubenswrapper[24928]: I1205 11:05:50.589195 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-logs\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.590467 master-0 kubenswrapper[24928]: I1205 11:05:50.589216 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-scripts\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.590467 master-0 kubenswrapper[24928]: I1205 11:05:50.589285 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ktrt\" (UniqueName: \"kubernetes.io/projected/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-kube-api-access-6ktrt\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.590467 master-0 kubenswrapper[24928]: I1205 11:05:50.589323 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-config-data\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.593241 master-0 kubenswrapper[24928]: I1205 11:05:50.592958 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-logs\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.594194 master-0 kubenswrapper[24928]: I1205 11:05:50.594133 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-scripts\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.594194 master-0 kubenswrapper[24928]: I1205 11:05:50.594147 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-httpd-run\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.596888 master-0 kubenswrapper[24928]: I1205 11:05:50.596517 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:05:50.596888 master-0 kubenswrapper[24928]: I1205 11:05:50.596570 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/4498c228462d6dd9d5d70d1a4aafdb52e778f9b7d49e65869f766f31c3adb188/globalmount\"" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.596888 master-0 kubenswrapper[24928]: I1205 11:05:50.596648 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-config-data\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.599466 master-0 kubenswrapper[24928]: I1205 11:05:50.598561 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-credential-keys\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.599586 master-0 kubenswrapper[24928]: I1205 11:05:50.599500 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-public-tls-certs\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.604495 master-0 kubenswrapper[24928]: I1205 11:05:50.600224 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-fernet-keys\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.604495 master-0 kubenswrapper[24928]: I1205 11:05:50.603984 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-scripts\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.605488 master-0 kubenswrapper[24928]: I1205 11:05:50.604956 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-combined-ca-bundle\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.605488 master-0 kubenswrapper[24928]: I1205 11:05:50.605126 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-config-data\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.612094 master-0 kubenswrapper[24928]: I1205 11:05:50.612024 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw8sl\" (UniqueName: \"kubernetes.io/projected/11a17674-7ccb-402d-9910-7ec8797f38b8-kube-api-access-vw8sl\") pod \"keystone-bootstrap-2wx5d\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:50.621606 master-0 kubenswrapper[24928]: I1205 11:05:50.616041 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-combined-ca-bundle\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.621606 master-0 kubenswrapper[24928]: I1205 11:05:50.618224 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ktrt\" (UniqueName: \"kubernetes.io/projected/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-kube-api-access-6ktrt\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:50.800565 master-0 kubenswrapper[24928]: I1205 11:05:50.800333 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:05:51.087861 master-0 kubenswrapper[24928]: I1205 11:05:51.087745 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-db-sync-sxns9"] Dec 05 11:05:51.089807 master-0 kubenswrapper[24928]: I1205 11:05:51.089710 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.092806 master-0 kubenswrapper[24928]: I1205 11:05:51.092675 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-config-data" Dec 05 11:05:51.093649 master-0 kubenswrapper[24928]: I1205 11:05:51.093611 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-scripts" Dec 05 11:05:51.200262 master-0 kubenswrapper[24928]: I1205 11:05:51.200174 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/0e18fdd2-be67-4228-82a3-4d02d3b350cf-etc-podinfo\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.201079 master-0 kubenswrapper[24928]: I1205 11:05:51.200318 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s48q2\" (UniqueName: \"kubernetes.io/projected/0e18fdd2-be67-4228-82a3-4d02d3b350cf-kube-api-access-s48q2\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.201079 master-0 kubenswrapper[24928]: I1205 11:05:51.200583 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-combined-ca-bundle\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.201079 master-0 kubenswrapper[24928]: I1205 11:05:51.200648 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.201261 master-0 kubenswrapper[24928]: I1205 11:05:51.201181 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-scripts\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.201416 master-0 kubenswrapper[24928]: I1205 11:05:51.201340 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data-merged\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.291534 master-0 kubenswrapper[24928]: I1205 11:05:51.291452 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vdmr2" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerName="registry-server" containerID="cri-o://32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e" gracePeriod=2 Dec 05 11:05:51.292259 master-0 kubenswrapper[24928]: I1205 11:05:51.291856 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-ec941-default-internal-api-0" podUID="c257c518-4db1-4ecd-abac-6bd4578c77d3" containerName="glance-log" containerID="cri-o://acab22f705229821317c8e656f6a30e41c6b91f42ae6d7d71598850b764f4626" gracePeriod=30 Dec 05 11:05:51.292259 master-0 kubenswrapper[24928]: I1205 11:05:51.291912 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"c257c518-4db1-4ecd-abac-6bd4578c77d3","Type":"ContainerStarted","Data":"362be91b94026357e8d2af037873328559cdbe4f8987853d7fb3ff44a50e162b"} Dec 05 11:05:51.292259 master-0 kubenswrapper[24928]: I1205 11:05:51.291983 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-ec941-default-internal-api-0" podUID="c257c518-4db1-4ecd-abac-6bd4578c77d3" containerName="glance-httpd" containerID="cri-o://362be91b94026357e8d2af037873328559cdbe4f8987853d7fb3ff44a50e162b" gracePeriod=30 Dec 05 11:05:51.312592 master-0 kubenswrapper[24928]: I1205 11:05:51.312120 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data-merged\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.313388 master-0 kubenswrapper[24928]: I1205 11:05:51.313275 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data-merged\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.314719 master-0 kubenswrapper[24928]: I1205 11:05:51.314641 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/0e18fdd2-be67-4228-82a3-4d02d3b350cf-etc-podinfo\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.315250 master-0 kubenswrapper[24928]: I1205 11:05:51.314892 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s48q2\" (UniqueName: \"kubernetes.io/projected/0e18fdd2-be67-4228-82a3-4d02d3b350cf-kube-api-access-s48q2\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.315250 master-0 kubenswrapper[24928]: I1205 11:05:51.315091 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-combined-ca-bundle\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.315250 master-0 kubenswrapper[24928]: I1205 11:05:51.315190 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.315995 master-0 kubenswrapper[24928]: I1205 11:05:51.315790 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-scripts\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.318047 master-0 kubenswrapper[24928]: I1205 11:05:51.317999 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/0e18fdd2-be67-4228-82a3-4d02d3b350cf-etc-podinfo\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.320307 master-0 kubenswrapper[24928]: I1205 11:05:51.320225 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-combined-ca-bundle\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.320515 master-0 kubenswrapper[24928]: I1205 11:05:51.320400 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.320983 master-0 kubenswrapper[24928]: I1205 11:05:51.320954 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-scripts\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:51.448348 master-0 kubenswrapper[24928]: I1205 11:05:51.448221 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-sync-sxns9"] Dec 05 11:05:52.026076 master-0 kubenswrapper[24928]: I1205 11:05:52.025980 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:52.222003 master-0 kubenswrapper[24928]: I1205 11:05:52.221850 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="577d038e-b40d-43e5-a6ad-ad4e278a3213" path="/var/lib/kubelet/pods/577d038e-b40d-43e5-a6ad-ad4e278a3213/volumes" Dec 05 11:05:52.222830 master-0 kubenswrapper[24928]: I1205 11:05:52.222788 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6140cfa3-11aa-4550-a8dc-ed81dcd087be" path="/var/lib/kubelet/pods/6140cfa3-11aa-4550-a8dc-ed81dcd087be/volumes" Dec 05 11:05:52.302806 master-0 kubenswrapper[24928]: I1205 11:05:52.302740 24928 generic.go:334] "Generic (PLEG): container finished" podID="c257c518-4db1-4ecd-abac-6bd4578c77d3" containerID="acab22f705229821317c8e656f6a30e41c6b91f42ae6d7d71598850b764f4626" exitCode=143 Dec 05 11:05:52.302806 master-0 kubenswrapper[24928]: I1205 11:05:52.302795 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"c257c518-4db1-4ecd-abac-6bd4578c77d3","Type":"ContainerDied","Data":"acab22f705229821317c8e656f6a30e41c6b91f42ae6d7d71598850b764f4626"} Dec 05 11:05:52.708021 master-0 kubenswrapper[24928]: I1205 11:05:52.707946 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s48q2\" (UniqueName: \"kubernetes.io/projected/0e18fdd2-be67-4228-82a3-4d02d3b350cf-kube-api-access-s48q2\") pod \"ironic-db-sync-sxns9\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:52.776153 master-0 kubenswrapper[24928]: I1205 11:05:52.776095 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2wx5d"] Dec 05 11:05:52.782495 master-0 kubenswrapper[24928]: I1205 11:05:52.780806 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:05:52.804133 master-0 kubenswrapper[24928]: I1205 11:05:52.804046 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-ec941-default-internal-api-0" podStartSLOduration=6.804027363 podStartE2EDuration="6.804027363s" podCreationTimestamp="2025-12-05 11:05:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:05:52.780282961 +0000 UTC m=+1112.783476812" watchObservedRunningTime="2025-12-05 11:05:52.804027363 +0000 UTC m=+1112.807221214" Dec 05 11:05:52.837452 master-0 kubenswrapper[24928]: I1205 11:05:52.835282 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:05:52.881079 master-0 kubenswrapper[24928]: I1205 11:05:52.880487 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5789dc4cf-qbmxl"] Dec 05 11:05:52.883660 master-0 kubenswrapper[24928]: I1205 11:05:52.881831 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" podUID="52a09227-8731-48e6-9b00-c38d799961f0" containerName="dnsmasq-dns" containerID="cri-o://83d7a70c489bb4eb9e4bb46585224d92019bcadd4ab5ad3fb37892a3a134feaa" gracePeriod=10 Dec 05 11:05:52.915513 master-0 kubenswrapper[24928]: I1205 11:05:52.914797 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-sxns9" Dec 05 11:05:53.053915 master-0 kubenswrapper[24928]: E1205 11:05:53.053863 24928 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98a4e639_cc2c_406e_b1b7_48746557cdc4.slice/crio-conmon-32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52a09227_8731_48e6_9b00_c38d799961f0.slice/crio-83d7a70c489bb4eb9e4bb46585224d92019bcadd4ab5ad3fb37892a3a134feaa.scope\": RecentStats: unable to find data in memory cache]" Dec 05 11:05:53.327481 master-0 kubenswrapper[24928]: I1205 11:05:53.327432 24928 generic.go:334] "Generic (PLEG): container finished" podID="c257c518-4db1-4ecd-abac-6bd4578c77d3" containerID="362be91b94026357e8d2af037873328559cdbe4f8987853d7fb3ff44a50e162b" exitCode=0 Dec 05 11:05:53.329314 master-0 kubenswrapper[24928]: I1205 11:05:53.328007 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"c257c518-4db1-4ecd-abac-6bd4578c77d3","Type":"ContainerDied","Data":"362be91b94026357e8d2af037873328559cdbe4f8987853d7fb3ff44a50e162b"} Dec 05 11:05:53.331046 master-0 kubenswrapper[24928]: I1205 11:05:53.330991 24928 generic.go:334] "Generic (PLEG): container finished" podID="cd047ad1-2e39-47f6-ad8f-97eb080e7766" containerID="0fd5eac4fad287c65188ecc90c8530f87659bfbe84a9ec67de0571e498c18967" exitCode=0 Dec 05 11:05:53.331144 master-0 kubenswrapper[24928]: I1205 11:05:53.331074 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-rrvhc" event={"ID":"cd047ad1-2e39-47f6-ad8f-97eb080e7766","Type":"ContainerDied","Data":"0fd5eac4fad287c65188ecc90c8530f87659bfbe84a9ec67de0571e498c18967"} Dec 05 11:05:53.337532 master-0 kubenswrapper[24928]: I1205 11:05:53.337490 24928 generic.go:334] "Generic (PLEG): container finished" podID="52a09227-8731-48e6-9b00-c38d799961f0" containerID="83d7a70c489bb4eb9e4bb46585224d92019bcadd4ab5ad3fb37892a3a134feaa" exitCode=0 Dec 05 11:05:53.337707 master-0 kubenswrapper[24928]: I1205 11:05:53.337578 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" event={"ID":"52a09227-8731-48e6-9b00-c38d799961f0","Type":"ContainerDied","Data":"83d7a70c489bb4eb9e4bb46585224d92019bcadd4ab5ad3fb37892a3a134feaa"} Dec 05 11:05:53.340795 master-0 kubenswrapper[24928]: I1205 11:05:53.340760 24928 generic.go:334] "Generic (PLEG): container finished" podID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerID="32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e" exitCode=0 Dec 05 11:05:53.340795 master-0 kubenswrapper[24928]: I1205 11:05:53.340796 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdmr2" event={"ID":"98a4e639-cc2c-406e-b1b7-48746557cdc4","Type":"ContainerDied","Data":"32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e"} Dec 05 11:05:56.822130 master-0 kubenswrapper[24928]: I1205 11:05:56.821894 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" podUID="52a09227-8731-48e6-9b00-c38d799961f0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.128.0.179:5353: connect: connection refused" Dec 05 11:05:59.921491 master-0 kubenswrapper[24928]: E1205 11:05:59.921387 24928 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e is running failed: container process not found" containerID="32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 11:05:59.922286 master-0 kubenswrapper[24928]: E1205 11:05:59.921926 24928 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e is running failed: container process not found" containerID="32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 11:05:59.922553 master-0 kubenswrapper[24928]: E1205 11:05:59.922382 24928 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e is running failed: container process not found" containerID="32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e" cmd=["grpc_health_probe","-addr=:50051"] Dec 05 11:05:59.922632 master-0 kubenswrapper[24928]: E1205 11:05:59.922585 24928 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-vdmr2" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerName="registry-server" Dec 05 11:06:01.822717 master-0 kubenswrapper[24928]: I1205 11:06:01.822562 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" podUID="52a09227-8731-48e6-9b00-c38d799961f0" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.128.0.179:5353: connect: connection refused" Dec 05 11:06:03.068454 master-0 kubenswrapper[24928]: I1205 11:06:03.067894 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:03.085552 master-0 kubenswrapper[24928]: I1205 11:06:03.085498 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-rrvhc" Dec 05 11:06:03.132578 master-0 kubenswrapper[24928]: I1205 11:06:03.132537 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:06:03.173316 master-0 kubenswrapper[24928]: I1205 11:06:03.172943 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-scripts\") pod \"c257c518-4db1-4ecd-abac-6bd4578c77d3\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " Dec 05 11:06:03.173628 master-0 kubenswrapper[24928]: I1205 11:06:03.173302 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"c257c518-4db1-4ecd-abac-6bd4578c77d3\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " Dec 05 11:06:03.173628 master-0 kubenswrapper[24928]: I1205 11:06:03.173516 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-combined-ca-bundle\") pod \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " Dec 05 11:06:03.173786 master-0 kubenswrapper[24928]: I1205 11:06:03.173714 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-scripts\") pod \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " Dec 05 11:06:03.173786 master-0 kubenswrapper[24928]: I1205 11:06:03.173781 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-combined-ca-bundle\") pod \"c257c518-4db1-4ecd-abac-6bd4578c77d3\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " Dec 05 11:06:03.173892 master-0 kubenswrapper[24928]: I1205 11:06:03.173851 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-config-data\") pod \"c257c518-4db1-4ecd-abac-6bd4578c77d3\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " Dec 05 11:06:03.173937 master-0 kubenswrapper[24928]: I1205 11:06:03.173915 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-config-data\") pod \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " Dec 05 11:06:03.174764 master-0 kubenswrapper[24928]: I1205 11:06:03.173954 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-httpd-run\") pod \"c257c518-4db1-4ecd-abac-6bd4578c77d3\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " Dec 05 11:06:03.174764 master-0 kubenswrapper[24928]: I1205 11:06:03.174087 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd047ad1-2e39-47f6-ad8f-97eb080e7766-logs\") pod \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " Dec 05 11:06:03.174764 master-0 kubenswrapper[24928]: I1205 11:06:03.174150 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-logs\") pod \"c257c518-4db1-4ecd-abac-6bd4578c77d3\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " Dec 05 11:06:03.174764 master-0 kubenswrapper[24928]: I1205 11:06:03.174246 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9w8c7\" (UniqueName: \"kubernetes.io/projected/c257c518-4db1-4ecd-abac-6bd4578c77d3-kube-api-access-9w8c7\") pod \"c257c518-4db1-4ecd-abac-6bd4578c77d3\" (UID: \"c257c518-4db1-4ecd-abac-6bd4578c77d3\") " Dec 05 11:06:03.174764 master-0 kubenswrapper[24928]: I1205 11:06:03.174327 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dp4q\" (UniqueName: \"kubernetes.io/projected/cd047ad1-2e39-47f6-ad8f-97eb080e7766-kube-api-access-5dp4q\") pod \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\" (UID: \"cd047ad1-2e39-47f6-ad8f-97eb080e7766\") " Dec 05 11:06:03.174937 master-0 kubenswrapper[24928]: I1205 11:06:03.174843 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c257c518-4db1-4ecd-abac-6bd4578c77d3" (UID: "c257c518-4db1-4ecd-abac-6bd4578c77d3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:06:03.175361 master-0 kubenswrapper[24928]: I1205 11:06:03.175326 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-logs" (OuterVolumeSpecName: "logs") pod "c257c518-4db1-4ecd-abac-6bd4578c77d3" (UID: "c257c518-4db1-4ecd-abac-6bd4578c77d3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:06:03.175545 master-0 kubenswrapper[24928]: I1205 11:06:03.175482 24928 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-httpd-run\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.175818 master-0 kubenswrapper[24928]: I1205 11:06:03.175795 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd047ad1-2e39-47f6-ad8f-97eb080e7766-logs" (OuterVolumeSpecName: "logs") pod "cd047ad1-2e39-47f6-ad8f-97eb080e7766" (UID: "cd047ad1-2e39-47f6-ad8f-97eb080e7766"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:06:03.181607 master-0 kubenswrapper[24928]: I1205 11:06:03.181544 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-scripts" (OuterVolumeSpecName: "scripts") pod "cd047ad1-2e39-47f6-ad8f-97eb080e7766" (UID: "cd047ad1-2e39-47f6-ad8f-97eb080e7766"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:03.181782 master-0 kubenswrapper[24928]: I1205 11:06:03.181601 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd047ad1-2e39-47f6-ad8f-97eb080e7766-kube-api-access-5dp4q" (OuterVolumeSpecName: "kube-api-access-5dp4q") pod "cd047ad1-2e39-47f6-ad8f-97eb080e7766" (UID: "cd047ad1-2e39-47f6-ad8f-97eb080e7766"). InnerVolumeSpecName "kube-api-access-5dp4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:03.181782 master-0 kubenswrapper[24928]: I1205 11:06:03.181619 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-scripts" (OuterVolumeSpecName: "scripts") pod "c257c518-4db1-4ecd-abac-6bd4578c77d3" (UID: "c257c518-4db1-4ecd-abac-6bd4578c77d3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:03.181959 master-0 kubenswrapper[24928]: I1205 11:06:03.181925 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c257c518-4db1-4ecd-abac-6bd4578c77d3-kube-api-access-9w8c7" (OuterVolumeSpecName: "kube-api-access-9w8c7") pod "c257c518-4db1-4ecd-abac-6bd4578c77d3" (UID: "c257c518-4db1-4ecd-abac-6bd4578c77d3"). InnerVolumeSpecName "kube-api-access-9w8c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:03.204217 master-0 kubenswrapper[24928]: I1205 11:06:03.204154 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:06:03.211201 master-0 kubenswrapper[24928]: I1205 11:06:03.211162 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880" (OuterVolumeSpecName: "glance") pod "c257c518-4db1-4ecd-abac-6bd4578c77d3" (UID: "c257c518-4db1-4ecd-abac-6bd4578c77d3"). InnerVolumeSpecName "pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 11:06:03.215280 master-0 kubenswrapper[24928]: I1205 11:06:03.215204 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c257c518-4db1-4ecd-abac-6bd4578c77d3" (UID: "c257c518-4db1-4ecd-abac-6bd4578c77d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:03.219375 master-0 kubenswrapper[24928]: I1205 11:06:03.219286 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cd047ad1-2e39-47f6-ad8f-97eb080e7766" (UID: "cd047ad1-2e39-47f6-ad8f-97eb080e7766"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:03.224323 master-0 kubenswrapper[24928]: I1205 11:06:03.224257 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-config-data" (OuterVolumeSpecName: "config-data") pod "cd047ad1-2e39-47f6-ad8f-97eb080e7766" (UID: "cd047ad1-2e39-47f6-ad8f-97eb080e7766"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:03.246779 master-0 kubenswrapper[24928]: I1205 11:06:03.246692 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-config-data" (OuterVolumeSpecName: "config-data") pod "c257c518-4db1-4ecd-abac-6bd4578c77d3" (UID: "c257c518-4db1-4ecd-abac-6bd4578c77d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:03.276713 master-0 kubenswrapper[24928]: I1205 11:06:03.276536 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-dns-svc\") pod \"52a09227-8731-48e6-9b00-c38d799961f0\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " Dec 05 11:06:03.276713 master-0 kubenswrapper[24928]: I1205 11:06:03.276627 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6l77\" (UniqueName: \"kubernetes.io/projected/52a09227-8731-48e6-9b00-c38d799961f0-kube-api-access-k6l77\") pod \"52a09227-8731-48e6-9b00-c38d799961f0\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " Dec 05 11:06:03.277001 master-0 kubenswrapper[24928]: I1205 11:06:03.276754 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfj6w\" (UniqueName: \"kubernetes.io/projected/98a4e639-cc2c-406e-b1b7-48746557cdc4-kube-api-access-cfj6w\") pod \"98a4e639-cc2c-406e-b1b7-48746557cdc4\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " Dec 05 11:06:03.277001 master-0 kubenswrapper[24928]: I1205 11:06:03.276806 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-nb\") pod \"52a09227-8731-48e6-9b00-c38d799961f0\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " Dec 05 11:06:03.277001 master-0 kubenswrapper[24928]: I1205 11:06:03.276898 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-catalog-content\") pod \"98a4e639-cc2c-406e-b1b7-48746557cdc4\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " Dec 05 11:06:03.277132 master-0 kubenswrapper[24928]: I1205 11:06:03.277011 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-utilities\") pod \"98a4e639-cc2c-406e-b1b7-48746557cdc4\" (UID: \"98a4e639-cc2c-406e-b1b7-48746557cdc4\") " Dec 05 11:06:03.277132 master-0 kubenswrapper[24928]: I1205 11:06:03.277071 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-config\") pod \"52a09227-8731-48e6-9b00-c38d799961f0\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " Dec 05 11:06:03.277132 master-0 kubenswrapper[24928]: I1205 11:06:03.277124 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-sb\") pod \"52a09227-8731-48e6-9b00-c38d799961f0\" (UID: \"52a09227-8731-48e6-9b00-c38d799961f0\") " Dec 05 11:06:03.277766 master-0 kubenswrapper[24928]: I1205 11:06:03.277725 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.277766 master-0 kubenswrapper[24928]: I1205 11:06:03.277757 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.277880 master-0 kubenswrapper[24928]: I1205 11:06:03.277769 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cd047ad1-2e39-47f6-ad8f-97eb080e7766-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.277880 master-0 kubenswrapper[24928]: I1205 11:06:03.277783 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c257c518-4db1-4ecd-abac-6bd4578c77d3-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.277880 master-0 kubenswrapper[24928]: I1205 11:06:03.277795 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9w8c7\" (UniqueName: \"kubernetes.io/projected/c257c518-4db1-4ecd-abac-6bd4578c77d3-kube-api-access-9w8c7\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.277880 master-0 kubenswrapper[24928]: I1205 11:06:03.277810 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dp4q\" (UniqueName: \"kubernetes.io/projected/cd047ad1-2e39-47f6-ad8f-97eb080e7766-kube-api-access-5dp4q\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.277880 master-0 kubenswrapper[24928]: I1205 11:06:03.277820 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.277880 master-0 kubenswrapper[24928]: I1205 11:06:03.277847 24928 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") on node \"master-0\" " Dec 05 11:06:03.277880 master-0 kubenswrapper[24928]: I1205 11:06:03.277862 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.277880 master-0 kubenswrapper[24928]: I1205 11:06:03.277873 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cd047ad1-2e39-47f6-ad8f-97eb080e7766-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.277880 master-0 kubenswrapper[24928]: I1205 11:06:03.277886 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c257c518-4db1-4ecd-abac-6bd4578c77d3-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.278240 master-0 kubenswrapper[24928]: I1205 11:06:03.278123 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-utilities" (OuterVolumeSpecName: "utilities") pod "98a4e639-cc2c-406e-b1b7-48746557cdc4" (UID: "98a4e639-cc2c-406e-b1b7-48746557cdc4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:06:03.284746 master-0 kubenswrapper[24928]: I1205 11:06:03.284689 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52a09227-8731-48e6-9b00-c38d799961f0-kube-api-access-k6l77" (OuterVolumeSpecName: "kube-api-access-k6l77") pod "52a09227-8731-48e6-9b00-c38d799961f0" (UID: "52a09227-8731-48e6-9b00-c38d799961f0"). InnerVolumeSpecName "kube-api-access-k6l77". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:03.285214 master-0 kubenswrapper[24928]: I1205 11:06:03.285159 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98a4e639-cc2c-406e-b1b7-48746557cdc4-kube-api-access-cfj6w" (OuterVolumeSpecName: "kube-api-access-cfj6w") pod "98a4e639-cc2c-406e-b1b7-48746557cdc4" (UID: "98a4e639-cc2c-406e-b1b7-48746557cdc4"). InnerVolumeSpecName "kube-api-access-cfj6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:03.498082 master-0 kubenswrapper[24928]: I1205 11:06:03.388744 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98a4e639-cc2c-406e-b1b7-48746557cdc4" (UID: "98a4e639-cc2c-406e-b1b7-48746557cdc4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:06:03.500534 master-0 kubenswrapper[24928]: I1205 11:06:03.499466 24928 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 11:06:03.500534 master-0 kubenswrapper[24928]: I1205 11:06:03.499714 24928 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01" (UniqueName: "kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880") on node "master-0" Dec 05 11:06:03.501929 master-0 kubenswrapper[24928]: I1205 11:06:03.501186 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6l77\" (UniqueName: \"kubernetes.io/projected/52a09227-8731-48e6-9b00-c38d799961f0-kube-api-access-k6l77\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.501929 master-0 kubenswrapper[24928]: I1205 11:06:03.501557 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfj6w\" (UniqueName: \"kubernetes.io/projected/98a4e639-cc2c-406e-b1b7-48746557cdc4-kube-api-access-cfj6w\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.501929 master-0 kubenswrapper[24928]: I1205 11:06:03.501580 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.501929 master-0 kubenswrapper[24928]: I1205 11:06:03.501594 24928 reconciler_common.go:293] "Volume detached for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.501929 master-0 kubenswrapper[24928]: I1205 11:06:03.501606 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98a4e639-cc2c-406e-b1b7-48746557cdc4-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.533812 master-0 kubenswrapper[24928]: I1205 11:06:03.533758 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "52a09227-8731-48e6-9b00-c38d799961f0" (UID: "52a09227-8731-48e6-9b00-c38d799961f0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:03.537450 master-0 kubenswrapper[24928]: I1205 11:06:03.536244 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "52a09227-8731-48e6-9b00-c38d799961f0" (UID: "52a09227-8731-48e6-9b00-c38d799961f0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:03.560039 master-0 kubenswrapper[24928]: I1205 11:06:03.559963 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "52a09227-8731-48e6-9b00-c38d799961f0" (UID: "52a09227-8731-48e6-9b00-c38d799961f0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:03.563026 master-0 kubenswrapper[24928]: I1205 11:06:03.562969 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-config" (OuterVolumeSpecName: "config") pod "52a09227-8731-48e6-9b00-c38d799961f0" (UID: "52a09227-8731-48e6-9b00-c38d799961f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:03.573469 master-0 kubenswrapper[24928]: I1205 11:06:03.573196 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vdmr2" Dec 05 11:06:03.573469 master-0 kubenswrapper[24928]: I1205 11:06:03.573231 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vdmr2" event={"ID":"98a4e639-cc2c-406e-b1b7-48746557cdc4","Type":"ContainerDied","Data":"da7afeabaace30849cec2c1dd38c2bdddf8fc79ff13b68dc39be74a112a8a416"} Dec 05 11:06:03.573469 master-0 kubenswrapper[24928]: I1205 11:06:03.573364 24928 scope.go:117] "RemoveContainer" containerID="32cc3f71178762a0601eb709be11fa93f03a4a1986e71762a3d65f409bf15f0e" Dec 05 11:06:03.578843 master-0 kubenswrapper[24928]: I1205 11:06:03.576810 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:03.578843 master-0 kubenswrapper[24928]: I1205 11:06:03.576843 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"c257c518-4db1-4ecd-abac-6bd4578c77d3","Type":"ContainerDied","Data":"845c2dd596934a91b79bd108e6d04ccfc513681cf6a760708c5de4c3208f38bd"} Dec 05 11:06:03.580606 master-0 kubenswrapper[24928]: I1205 11:06:03.580549 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-rrvhc" event={"ID":"cd047ad1-2e39-47f6-ad8f-97eb080e7766","Type":"ContainerDied","Data":"706172a6d379ad38ce7938c7ff6bac1f92546191b7fd0dc5fe68bae400a24cf4"} Dec 05 11:06:03.580606 master-0 kubenswrapper[24928]: I1205 11:06:03.580595 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="706172a6d379ad38ce7938c7ff6bac1f92546191b7fd0dc5fe68bae400a24cf4" Dec 05 11:06:03.580606 master-0 kubenswrapper[24928]: I1205 11:06:03.580602 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-rrvhc" Dec 05 11:06:03.588168 master-0 kubenswrapper[24928]: I1205 11:06:03.588099 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" event={"ID":"52a09227-8731-48e6-9b00-c38d799961f0","Type":"ContainerDied","Data":"592a04a817e5713d439060d1e1e8b48e0127ea7799ca0b4415272c74d43ecc32"} Dec 05 11:06:03.588168 master-0 kubenswrapper[24928]: I1205 11:06:03.588158 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5789dc4cf-qbmxl" Dec 05 11:06:03.590784 master-0 kubenswrapper[24928]: I1205 11:06:03.590734 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2wx5d" event={"ID":"11a17674-7ccb-402d-9910-7ec8797f38b8","Type":"ContainerStarted","Data":"612a89b78524e8e7ec8f1b6f48f4864f9f31a47c6deaf2794d8e41c671c1fc6f"} Dec 05 11:06:03.590880 master-0 kubenswrapper[24928]: I1205 11:06:03.590799 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2wx5d" event={"ID":"11a17674-7ccb-402d-9910-7ec8797f38b8","Type":"ContainerStarted","Data":"b00e009bd22a5d2789544d34bc48ce687970dddfb86f2c68db745b5224f35c81"} Dec 05 11:06:03.595766 master-0 kubenswrapper[24928]: I1205 11:06:03.595720 24928 scope.go:117] "RemoveContainer" containerID="465eaaacdeaddc9af9ecd8ff210c679f3a485340cb493223b88c974544b2b75d" Dec 05 11:06:03.604434 master-0 kubenswrapper[24928]: I1205 11:06:03.604365 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.604434 master-0 kubenswrapper[24928]: I1205 11:06:03.604410 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.604593 master-0 kubenswrapper[24928]: I1205 11:06:03.604452 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-ovsdbserver-sb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.604593 master-0 kubenswrapper[24928]: I1205 11:06:03.604475 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52a09227-8731-48e6-9b00-c38d799961f0-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:03.636292 master-0 kubenswrapper[24928]: I1205 11:06:03.636236 24928 scope.go:117] "RemoveContainer" containerID="40ffa0c10cc39cb58eb889575620e96ab5d9b9a2153e361fe6094e4d63c6ff83" Dec 05 11:06:03.665621 master-0 kubenswrapper[24928]: I1205 11:06:03.665552 24928 scope.go:117] "RemoveContainer" containerID="362be91b94026357e8d2af037873328559cdbe4f8987853d7fb3ff44a50e162b" Dec 05 11:06:03.691158 master-0 kubenswrapper[24928]: I1205 11:06:03.691083 24928 scope.go:117] "RemoveContainer" containerID="acab22f705229821317c8e656f6a30e41c6b91f42ae6d7d71598850b764f4626" Dec 05 11:06:03.711450 master-0 kubenswrapper[24928]: I1205 11:06:03.711278 24928 scope.go:117] "RemoveContainer" containerID="83d7a70c489bb4eb9e4bb46585224d92019bcadd4ab5ad3fb37892a3a134feaa" Dec 05 11:06:03.747936 master-0 kubenswrapper[24928]: I1205 11:06:03.747736 24928 scope.go:117] "RemoveContainer" containerID="f8a129e1c3ea5111b73e5f09a16e2e59cc6ca1917594df247773587916a7a399" Dec 05 11:06:04.294455 master-0 kubenswrapper[24928]: I1205 11:06:04.294361 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-sync-sxns9"] Dec 05 11:06:04.618755 master-0 kubenswrapper[24928]: I1205 11:06:04.618658 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-db-sync-9nbmc" event={"ID":"4ae72689-6505-4064-bd26-861bda2f68cc","Type":"ContainerStarted","Data":"349e2e75b98f5ec0c4eb43ce361e97dfbc1848d82e597f7be07ffc8be044fa44"} Dec 05 11:06:04.622622 master-0 kubenswrapper[24928]: I1205 11:06:04.622518 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-sxns9" event={"ID":"0e18fdd2-be67-4228-82a3-4d02d3b350cf","Type":"ContainerStarted","Data":"d3a679d6a69ae226505b1d8a46dfc2aff149e33ae9d32741983f6be8bd090375"} Dec 05 11:06:05.538673 master-0 kubenswrapper[24928]: I1205 11:06:05.532625 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5789dc4cf-qbmxl"] Dec 05 11:06:05.644558 master-0 kubenswrapper[24928]: I1205 11:06:05.644474 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5789dc4cf-qbmxl"] Dec 05 11:06:05.809471 master-0 kubenswrapper[24928]: W1205 11:06:05.809369 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1a0529b_71c2_4a72_b75f_0b56b07e7c64.slice/crio-e091c68b96876c484a694f5b55438623d568ae6bf609dd6717299b74ac97cc8e WatchSource:0}: Error finding container e091c68b96876c484a694f5b55438623d568ae6bf609dd6717299b74ac97cc8e: Status 404 returned error can't find the container with id e091c68b96876c484a694f5b55438623d568ae6bf609dd6717299b74ac97cc8e Dec 05 11:06:05.823263 master-0 kubenswrapper[24928]: I1205 11:06:05.817458 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:06:05.969066 master-0 kubenswrapper[24928]: I1205 11:06:05.968986 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vdmr2"] Dec 05 11:06:05.971446 master-0 kubenswrapper[24928]: I1205 11:06:05.971388 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vdmr2"] Dec 05 11:06:05.984916 master-0 kubenswrapper[24928]: I1205 11:06:05.984809 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:06:05.997333 master-0 kubenswrapper[24928]: I1205 11:06:05.997223 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:06:05.998687 master-0 kubenswrapper[24928]: I1205 11:06:05.998558 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-ec941-db-sync-9nbmc" podStartSLOduration=4.972957245 podStartE2EDuration="25.998533528s" podCreationTimestamp="2025-12-05 11:05:40 +0000 UTC" firstStartedPulling="2025-12-05 11:05:41.903358276 +0000 UTC m=+1101.906552127" lastFinishedPulling="2025-12-05 11:06:02.928934559 +0000 UTC m=+1122.932128410" observedRunningTime="2025-12-05 11:06:05.896275808 +0000 UTC m=+1125.899469669" watchObservedRunningTime="2025-12-05 11:06:05.998533528 +0000 UTC m=+1126.001727389" Dec 05 11:06:06.025463 master-0 kubenswrapper[24928]: I1205 11:06:06.025377 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:06:06.026006 master-0 kubenswrapper[24928]: E1205 11:06:06.025972 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a09227-8731-48e6-9b00-c38d799961f0" containerName="dnsmasq-dns" Dec 05 11:06:06.026006 master-0 kubenswrapper[24928]: I1205 11:06:06.025996 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a09227-8731-48e6-9b00-c38d799961f0" containerName="dnsmasq-dns" Dec 05 11:06:06.026104 master-0 kubenswrapper[24928]: E1205 11:06:06.026014 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerName="extract-content" Dec 05 11:06:06.026104 master-0 kubenswrapper[24928]: I1205 11:06:06.026022 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerName="extract-content" Dec 05 11:06:06.026104 master-0 kubenswrapper[24928]: E1205 11:06:06.026041 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerName="extract-utilities" Dec 05 11:06:06.026104 master-0 kubenswrapper[24928]: I1205 11:06:06.026048 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerName="extract-utilities" Dec 05 11:06:06.026104 master-0 kubenswrapper[24928]: E1205 11:06:06.026064 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52a09227-8731-48e6-9b00-c38d799961f0" containerName="init" Dec 05 11:06:06.026104 master-0 kubenswrapper[24928]: I1205 11:06:06.026070 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="52a09227-8731-48e6-9b00-c38d799961f0" containerName="init" Dec 05 11:06:06.026284 master-0 kubenswrapper[24928]: E1205 11:06:06.026112 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd047ad1-2e39-47f6-ad8f-97eb080e7766" containerName="placement-db-sync" Dec 05 11:06:06.026284 master-0 kubenswrapper[24928]: I1205 11:06:06.026119 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd047ad1-2e39-47f6-ad8f-97eb080e7766" containerName="placement-db-sync" Dec 05 11:06:06.026284 master-0 kubenswrapper[24928]: E1205 11:06:06.026141 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c257c518-4db1-4ecd-abac-6bd4578c77d3" containerName="glance-httpd" Dec 05 11:06:06.026284 master-0 kubenswrapper[24928]: I1205 11:06:06.026147 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c257c518-4db1-4ecd-abac-6bd4578c77d3" containerName="glance-httpd" Dec 05 11:06:06.026284 master-0 kubenswrapper[24928]: E1205 11:06:06.026164 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerName="registry-server" Dec 05 11:06:06.026284 master-0 kubenswrapper[24928]: I1205 11:06:06.026170 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerName="registry-server" Dec 05 11:06:06.026284 master-0 kubenswrapper[24928]: E1205 11:06:06.026181 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c257c518-4db1-4ecd-abac-6bd4578c77d3" containerName="glance-log" Dec 05 11:06:06.026284 master-0 kubenswrapper[24928]: I1205 11:06:06.026187 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c257c518-4db1-4ecd-abac-6bd4578c77d3" containerName="glance-log" Dec 05 11:06:06.026863 master-0 kubenswrapper[24928]: I1205 11:06:06.026825 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd047ad1-2e39-47f6-ad8f-97eb080e7766" containerName="placement-db-sync" Dec 05 11:06:06.026863 master-0 kubenswrapper[24928]: I1205 11:06:06.026857 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" containerName="registry-server" Dec 05 11:06:06.027049 master-0 kubenswrapper[24928]: I1205 11:06:06.026881 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="52a09227-8731-48e6-9b00-c38d799961f0" containerName="dnsmasq-dns" Dec 05 11:06:06.027049 master-0 kubenswrapper[24928]: I1205 11:06:06.026912 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c257c518-4db1-4ecd-abac-6bd4578c77d3" containerName="glance-httpd" Dec 05 11:06:06.027049 master-0 kubenswrapper[24928]: I1205 11:06:06.026938 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c257c518-4db1-4ecd-abac-6bd4578c77d3" containerName="glance-log" Dec 05 11:06:06.028789 master-0 kubenswrapper[24928]: I1205 11:06:06.028731 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.038451 master-0 kubenswrapper[24928]: I1205 11:06:06.038163 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:06:06.038720 master-0 kubenswrapper[24928]: I1205 11:06:06.038515 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2wx5d" podStartSLOduration=16.038502045 podStartE2EDuration="16.038502045s" podCreationTimestamp="2025-12-05 11:05:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:05.96895968 +0000 UTC m=+1125.972153531" watchObservedRunningTime="2025-12-05 11:06:06.038502045 +0000 UTC m=+1126.041695906" Dec 05 11:06:06.039842 master-0 kubenswrapper[24928]: I1205 11:06:06.039795 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 11:06:06.040213 master-0 kubenswrapper[24928]: I1205 11:06:06.040144 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-ec941-default-internal-config-data" Dec 05 11:06:06.075742 master-0 kubenswrapper[24928]: I1205 11:06:06.075539 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.075742 master-0 kubenswrapper[24928]: I1205 11:06:06.075628 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdjll\" (UniqueName: \"kubernetes.io/projected/fddf0e6c-7977-4429-9015-732fd8dbb6a1-kube-api-access-bdjll\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.075742 master-0 kubenswrapper[24928]: I1205 11:06:06.075655 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-internal-tls-certs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.075742 master-0 kubenswrapper[24928]: I1205 11:06:06.075689 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.075742 master-0 kubenswrapper[24928]: I1205 11:06:06.075729 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.076210 master-0 kubenswrapper[24928]: I1205 11:06:06.075889 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.076210 master-0 kubenswrapper[24928]: I1205 11:06:06.076028 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.076210 master-0 kubenswrapper[24928]: I1205 11:06:06.076092 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.178513 master-0 kubenswrapper[24928]: I1205 11:06:06.178403 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.178754 master-0 kubenswrapper[24928]: I1205 11:06:06.178523 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.178754 master-0 kubenswrapper[24928]: I1205 11:06:06.178606 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.178754 master-0 kubenswrapper[24928]: I1205 11:06:06.178642 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdjll\" (UniqueName: \"kubernetes.io/projected/fddf0e6c-7977-4429-9015-732fd8dbb6a1-kube-api-access-bdjll\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.178754 master-0 kubenswrapper[24928]: I1205 11:06:06.178669 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-internal-tls-certs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.178754 master-0 kubenswrapper[24928]: I1205 11:06:06.178706 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.178754 master-0 kubenswrapper[24928]: I1205 11:06:06.178742 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.178953 master-0 kubenswrapper[24928]: I1205 11:06:06.178786 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.182581 master-0 kubenswrapper[24928]: I1205 11:06:06.182258 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.182814 master-0 kubenswrapper[24928]: I1205 11:06:06.182736 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.186483 master-0 kubenswrapper[24928]: I1205 11:06:06.186400 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:06:06.186679 master-0 kubenswrapper[24928]: I1205 11:06:06.186496 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/f829e9071eee306944ccf41eabbcf74a1473d7031ba724c821f61b9a30a289e9/globalmount\"" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.193668 master-0 kubenswrapper[24928]: I1205 11:06:06.190677 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.193668 master-0 kubenswrapper[24928]: I1205 11:06:06.191879 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.193668 master-0 kubenswrapper[24928]: I1205 11:06:06.193134 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-internal-tls-certs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.216830 master-0 kubenswrapper[24928]: I1205 11:06:06.206598 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.216830 master-0 kubenswrapper[24928]: I1205 11:06:06.210471 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdjll\" (UniqueName: \"kubernetes.io/projected/fddf0e6c-7977-4429-9015-732fd8dbb6a1-kube-api-access-bdjll\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:06.228646 master-0 kubenswrapper[24928]: I1205 11:06:06.228534 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52a09227-8731-48e6-9b00-c38d799961f0" path="/var/lib/kubelet/pods/52a09227-8731-48e6-9b00-c38d799961f0/volumes" Dec 05 11:06:06.229456 master-0 kubenswrapper[24928]: I1205 11:06:06.229413 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98a4e639-cc2c-406e-b1b7-48746557cdc4" path="/var/lib/kubelet/pods/98a4e639-cc2c-406e-b1b7-48746557cdc4/volumes" Dec 05 11:06:06.230259 master-0 kubenswrapper[24928]: I1205 11:06:06.230225 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c257c518-4db1-4ecd-abac-6bd4578c77d3" path="/var/lib/kubelet/pods/c257c518-4db1-4ecd-abac-6bd4578c77d3/volumes" Dec 05 11:06:06.444560 master-0 kubenswrapper[24928]: I1205 11:06:06.443861 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6dc4d774b6-b9nzv"] Dec 05 11:06:06.451185 master-0 kubenswrapper[24928]: I1205 11:06:06.447227 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.458005 master-0 kubenswrapper[24928]: I1205 11:06:06.452978 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 05 11:06:06.458005 master-0 kubenswrapper[24928]: I1205 11:06:06.453248 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 11:06:06.458005 master-0 kubenswrapper[24928]: I1205 11:06:06.453404 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 11:06:06.458005 master-0 kubenswrapper[24928]: I1205 11:06:06.453678 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 05 11:06:06.458005 master-0 kubenswrapper[24928]: I1205 11:06:06.456987 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6dc4d774b6-b9nzv"] Dec 05 11:06:06.485889 master-0 kubenswrapper[24928]: I1205 11:06:06.485829 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-internal-tls-certs\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.486146 master-0 kubenswrapper[24928]: I1205 11:06:06.486118 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-scripts\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.486335 master-0 kubenswrapper[24928]: I1205 11:06:06.486312 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-combined-ca-bundle\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.486623 master-0 kubenswrapper[24928]: I1205 11:06:06.486600 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/091c2417-b57e-4402-acaa-53853e98de73-logs\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.486752 master-0 kubenswrapper[24928]: I1205 11:06:06.486732 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqkf7\" (UniqueName: \"kubernetes.io/projected/091c2417-b57e-4402-acaa-53853e98de73-kube-api-access-lqkf7\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.486888 master-0 kubenswrapper[24928]: I1205 11:06:06.486867 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-public-tls-certs\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.487048 master-0 kubenswrapper[24928]: I1205 11:06:06.487030 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-config-data\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.590100 master-0 kubenswrapper[24928]: I1205 11:06:06.589935 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-combined-ca-bundle\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.595061 master-0 kubenswrapper[24928]: I1205 11:06:06.590116 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/091c2417-b57e-4402-acaa-53853e98de73-logs\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.595061 master-0 kubenswrapper[24928]: I1205 11:06:06.590161 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqkf7\" (UniqueName: \"kubernetes.io/projected/091c2417-b57e-4402-acaa-53853e98de73-kube-api-access-lqkf7\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.595061 master-0 kubenswrapper[24928]: I1205 11:06:06.590212 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-public-tls-certs\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.595061 master-0 kubenswrapper[24928]: I1205 11:06:06.590293 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-config-data\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.595061 master-0 kubenswrapper[24928]: I1205 11:06:06.590722 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-internal-tls-certs\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.595061 master-0 kubenswrapper[24928]: I1205 11:06:06.590938 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-scripts\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.595061 master-0 kubenswrapper[24928]: I1205 11:06:06.591466 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/091c2417-b57e-4402-acaa-53853e98de73-logs\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.597843 master-0 kubenswrapper[24928]: I1205 11:06:06.597607 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-scripts\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.599239 master-0 kubenswrapper[24928]: I1205 11:06:06.599167 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-config-data\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.599512 master-0 kubenswrapper[24928]: I1205 11:06:06.599451 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-combined-ca-bundle\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.614073 master-0 kubenswrapper[24928]: I1205 11:06:06.605151 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-internal-tls-certs\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.614829 master-0 kubenswrapper[24928]: I1205 11:06:06.614517 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/091c2417-b57e-4402-acaa-53853e98de73-public-tls-certs\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.624347 master-0 kubenswrapper[24928]: I1205 11:06:06.624278 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqkf7\" (UniqueName: \"kubernetes.io/projected/091c2417-b57e-4402-acaa-53853e98de73-kube-api-access-lqkf7\") pod \"placement-6dc4d774b6-b9nzv\" (UID: \"091c2417-b57e-4402-acaa-53853e98de73\") " pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:06.672152 master-0 kubenswrapper[24928]: I1205 11:06:06.672003 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"e1a0529b-71c2-4a72-b75f-0b56b07e7c64","Type":"ContainerStarted","Data":"73bb8e24a58ffbe2967e4ba93ebbac3bff4baac27f7243d2c461aaa0802ea5d2"} Dec 05 11:06:06.672152 master-0 kubenswrapper[24928]: I1205 11:06:06.672076 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"e1a0529b-71c2-4a72-b75f-0b56b07e7c64","Type":"ContainerStarted","Data":"e091c68b96876c484a694f5b55438623d568ae6bf609dd6717299b74ac97cc8e"} Dec 05 11:06:06.785444 master-0 kubenswrapper[24928]: I1205 11:06:06.784968 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:07.324159 master-0 kubenswrapper[24928]: W1205 11:06:07.324088 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod091c2417_b57e_4402_acaa_53853e98de73.slice/crio-8c8d861a11af8f03ec5e4661aab14c248842a9b5999c07261dc8757d80d6703d WatchSource:0}: Error finding container 8c8d861a11af8f03ec5e4661aab14c248842a9b5999c07261dc8757d80d6703d: Status 404 returned error can't find the container with id 8c8d861a11af8f03ec5e4661aab14c248842a9b5999c07261dc8757d80d6703d Dec 05 11:06:07.326142 master-0 kubenswrapper[24928]: I1205 11:06:07.326074 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6dc4d774b6-b9nzv"] Dec 05 11:06:07.631101 master-0 kubenswrapper[24928]: I1205 11:06:07.630298 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:07.698988 master-0 kubenswrapper[24928]: I1205 11:06:07.698866 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"e1a0529b-71c2-4a72-b75f-0b56b07e7c64","Type":"ContainerStarted","Data":"acd53d82e3153e795b5309b7c86275854aea6e0fe04723a35820bec9073a5ad8"} Dec 05 11:06:07.703780 master-0 kubenswrapper[24928]: I1205 11:06:07.703683 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dc4d774b6-b9nzv" event={"ID":"091c2417-b57e-4402-acaa-53853e98de73","Type":"ContainerStarted","Data":"89a3ec13d8148c0803c967bc2cf65f14a803ee89e1067830d8a21eaadce43cf5"} Dec 05 11:06:07.703780 master-0 kubenswrapper[24928]: I1205 11:06:07.703763 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dc4d774b6-b9nzv" event={"ID":"091c2417-b57e-4402-acaa-53853e98de73","Type":"ContainerStarted","Data":"8c8d861a11af8f03ec5e4661aab14c248842a9b5999c07261dc8757d80d6703d"} Dec 05 11:06:07.708361 master-0 kubenswrapper[24928]: I1205 11:06:07.708293 24928 generic.go:334] "Generic (PLEG): container finished" podID="11a17674-7ccb-402d-9910-7ec8797f38b8" containerID="612a89b78524e8e7ec8f1b6f48f4864f9f31a47c6deaf2794d8e41c671c1fc6f" exitCode=0 Dec 05 11:06:07.708361 master-0 kubenswrapper[24928]: I1205 11:06:07.708352 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2wx5d" event={"ID":"11a17674-7ccb-402d-9910-7ec8797f38b8","Type":"ContainerDied","Data":"612a89b78524e8e7ec8f1b6f48f4864f9f31a47c6deaf2794d8e41c671c1fc6f"} Dec 05 11:06:07.742503 master-0 kubenswrapper[24928]: I1205 11:06:07.730748 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-ec941-default-external-api-0" podStartSLOduration=17.730716181 podStartE2EDuration="17.730716181s" podCreationTimestamp="2025-12-05 11:05:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:07.729848989 +0000 UTC m=+1127.733042860" watchObservedRunningTime="2025-12-05 11:06:07.730716181 +0000 UTC m=+1127.733910032" Dec 05 11:06:07.876574 master-0 kubenswrapper[24928]: I1205 11:06:07.876402 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:09.740104 master-0 kubenswrapper[24928]: I1205 11:06:09.740041 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:06:09.745115 master-0 kubenswrapper[24928]: I1205 11:06:09.745049 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2wx5d" event={"ID":"11a17674-7ccb-402d-9910-7ec8797f38b8","Type":"ContainerDied","Data":"b00e009bd22a5d2789544d34bc48ce687970dddfb86f2c68db745b5224f35c81"} Dec 05 11:06:09.745115 master-0 kubenswrapper[24928]: I1205 11:06:09.745101 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2wx5d" Dec 05 11:06:09.745621 master-0 kubenswrapper[24928]: I1205 11:06:09.745107 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b00e009bd22a5d2789544d34bc48ce687970dddfb86f2c68db745b5224f35c81" Dec 05 11:06:09.782006 master-0 kubenswrapper[24928]: I1205 11:06:09.777616 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-config-data\") pod \"11a17674-7ccb-402d-9910-7ec8797f38b8\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " Dec 05 11:06:09.782006 master-0 kubenswrapper[24928]: I1205 11:06:09.777796 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-combined-ca-bundle\") pod \"11a17674-7ccb-402d-9910-7ec8797f38b8\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " Dec 05 11:06:09.782006 master-0 kubenswrapper[24928]: I1205 11:06:09.777868 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-credential-keys\") pod \"11a17674-7ccb-402d-9910-7ec8797f38b8\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " Dec 05 11:06:09.782006 master-0 kubenswrapper[24928]: I1205 11:06:09.777933 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vw8sl\" (UniqueName: \"kubernetes.io/projected/11a17674-7ccb-402d-9910-7ec8797f38b8-kube-api-access-vw8sl\") pod \"11a17674-7ccb-402d-9910-7ec8797f38b8\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " Dec 05 11:06:09.782006 master-0 kubenswrapper[24928]: I1205 11:06:09.777984 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-fernet-keys\") pod \"11a17674-7ccb-402d-9910-7ec8797f38b8\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " Dec 05 11:06:09.782006 master-0 kubenswrapper[24928]: I1205 11:06:09.778036 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-scripts\") pod \"11a17674-7ccb-402d-9910-7ec8797f38b8\" (UID: \"11a17674-7ccb-402d-9910-7ec8797f38b8\") " Dec 05 11:06:09.782932 master-0 kubenswrapper[24928]: I1205 11:06:09.782739 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-scripts" (OuterVolumeSpecName: "scripts") pod "11a17674-7ccb-402d-9910-7ec8797f38b8" (UID: "11a17674-7ccb-402d-9910-7ec8797f38b8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:09.790224 master-0 kubenswrapper[24928]: I1205 11:06:09.783741 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "11a17674-7ccb-402d-9910-7ec8797f38b8" (UID: "11a17674-7ccb-402d-9910-7ec8797f38b8"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:09.790224 master-0 kubenswrapper[24928]: I1205 11:06:09.787914 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "11a17674-7ccb-402d-9910-7ec8797f38b8" (UID: "11a17674-7ccb-402d-9910-7ec8797f38b8"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:09.795632 master-0 kubenswrapper[24928]: I1205 11:06:09.795354 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11a17674-7ccb-402d-9910-7ec8797f38b8-kube-api-access-vw8sl" (OuterVolumeSpecName: "kube-api-access-vw8sl") pod "11a17674-7ccb-402d-9910-7ec8797f38b8" (UID: "11a17674-7ccb-402d-9910-7ec8797f38b8"). InnerVolumeSpecName "kube-api-access-vw8sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:09.837567 master-0 kubenswrapper[24928]: I1205 11:06:09.837492 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-config-data" (OuterVolumeSpecName: "config-data") pod "11a17674-7ccb-402d-9910-7ec8797f38b8" (UID: "11a17674-7ccb-402d-9910-7ec8797f38b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:09.844155 master-0 kubenswrapper[24928]: I1205 11:06:09.844090 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11a17674-7ccb-402d-9910-7ec8797f38b8" (UID: "11a17674-7ccb-402d-9910-7ec8797f38b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:09.881035 master-0 kubenswrapper[24928]: I1205 11:06:09.880887 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:09.881035 master-0 kubenswrapper[24928]: I1205 11:06:09.880962 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:09.881035 master-0 kubenswrapper[24928]: I1205 11:06:09.880981 24928 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-credential-keys\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:09.881035 master-0 kubenswrapper[24928]: I1205 11:06:09.880999 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vw8sl\" (UniqueName: \"kubernetes.io/projected/11a17674-7ccb-402d-9910-7ec8797f38b8-kube-api-access-vw8sl\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:09.881035 master-0 kubenswrapper[24928]: I1205 11:06:09.881013 24928 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-fernet-keys\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:09.881035 master-0 kubenswrapper[24928]: I1205 11:06:09.881024 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11a17674-7ccb-402d-9910-7ec8797f38b8-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:12.010474 master-0 kubenswrapper[24928]: I1205 11:06:12.010385 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-558cb7d6b-wgm2l"] Dec 05 11:06:12.011161 master-0 kubenswrapper[24928]: E1205 11:06:12.010881 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11a17674-7ccb-402d-9910-7ec8797f38b8" containerName="keystone-bootstrap" Dec 05 11:06:12.011161 master-0 kubenswrapper[24928]: I1205 11:06:12.010902 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="11a17674-7ccb-402d-9910-7ec8797f38b8" containerName="keystone-bootstrap" Dec 05 11:06:12.011293 master-0 kubenswrapper[24928]: I1205 11:06:12.011179 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="11a17674-7ccb-402d-9910-7ec8797f38b8" containerName="keystone-bootstrap" Dec 05 11:06:12.012010 master-0 kubenswrapper[24928]: I1205 11:06:12.011967 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.014096 master-0 kubenswrapper[24928]: I1205 11:06:12.014055 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 05 11:06:12.014440 master-0 kubenswrapper[24928]: I1205 11:06:12.014385 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 11:06:12.015050 master-0 kubenswrapper[24928]: I1205 11:06:12.015015 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 11:06:12.015369 master-0 kubenswrapper[24928]: I1205 11:06:12.015315 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 05 11:06:12.015583 master-0 kubenswrapper[24928]: I1205 11:06:12.015557 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 11:06:12.127603 master-0 kubenswrapper[24928]: I1205 11:06:12.127532 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjpx8\" (UniqueName: \"kubernetes.io/projected/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-kube-api-access-rjpx8\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.127786 master-0 kubenswrapper[24928]: I1205 11:06:12.127641 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-internal-tls-certs\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.127915 master-0 kubenswrapper[24928]: I1205 11:06:12.127857 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-fernet-keys\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.128093 master-0 kubenswrapper[24928]: I1205 11:06:12.128056 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-credential-keys\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.128278 master-0 kubenswrapper[24928]: I1205 11:06:12.128249 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-config-data\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.128381 master-0 kubenswrapper[24928]: I1205 11:06:12.128355 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-scripts\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.128472 master-0 kubenswrapper[24928]: I1205 11:06:12.128442 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-public-tls-certs\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.128587 master-0 kubenswrapper[24928]: I1205 11:06:12.128559 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-combined-ca-bundle\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.246013 master-0 kubenswrapper[24928]: I1205 11:06:12.231347 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-config-data\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.246013 master-0 kubenswrapper[24928]: I1205 11:06:12.231479 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-scripts\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.246013 master-0 kubenswrapper[24928]: I1205 11:06:12.231525 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-public-tls-certs\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.246013 master-0 kubenswrapper[24928]: I1205 11:06:12.238561 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-combined-ca-bundle\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.246013 master-0 kubenswrapper[24928]: I1205 11:06:12.238687 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjpx8\" (UniqueName: \"kubernetes.io/projected/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-kube-api-access-rjpx8\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.246013 master-0 kubenswrapper[24928]: I1205 11:06:12.238781 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-internal-tls-certs\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.246013 master-0 kubenswrapper[24928]: I1205 11:06:12.238906 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-fernet-keys\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.246013 master-0 kubenswrapper[24928]: I1205 11:06:12.239046 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-credential-keys\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.256971 master-0 kubenswrapper[24928]: I1205 11:06:12.254076 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-credential-keys\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.256971 master-0 kubenswrapper[24928]: I1205 11:06:12.254819 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-558cb7d6b-wgm2l"] Dec 05 11:06:12.256971 master-0 kubenswrapper[24928]: I1205 11:06:12.255989 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-scripts\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.259308 master-0 kubenswrapper[24928]: I1205 11:06:12.259261 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-config-data\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.259417 master-0 kubenswrapper[24928]: I1205 11:06:12.259266 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-fernet-keys\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.259713 master-0 kubenswrapper[24928]: I1205 11:06:12.259671 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-internal-tls-certs\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.259773 master-0 kubenswrapper[24928]: I1205 11:06:12.259719 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-public-tls-certs\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.270166 master-0 kubenswrapper[24928]: I1205 11:06:12.270025 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-combined-ca-bundle\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:12.784169 master-0 kubenswrapper[24928]: I1205 11:06:12.784103 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6dc4d774b6-b9nzv" event={"ID":"091c2417-b57e-4402-acaa-53853e98de73","Type":"ContainerStarted","Data":"b94e2d919927fd23bb8ddd9ac4c797ec7fa894ac85a85dba5e8131db8fd4d809"} Dec 05 11:06:12.784620 master-0 kubenswrapper[24928]: I1205 11:06:12.784583 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:12.836789 master-0 kubenswrapper[24928]: I1205 11:06:12.836697 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:06:12.836789 master-0 kubenswrapper[24928]: I1205 11:06:12.836770 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:06:12.870072 master-0 kubenswrapper[24928]: I1205 11:06:12.869977 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:06:12.879998 master-0 kubenswrapper[24928]: I1205 11:06:12.879926 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:06:13.446009 master-0 kubenswrapper[24928]: I1205 11:06:13.445953 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjpx8\" (UniqueName: \"kubernetes.io/projected/1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5-kube-api-access-rjpx8\") pod \"keystone-558cb7d6b-wgm2l\" (UID: \"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5\") " pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:13.532228 master-0 kubenswrapper[24928]: I1205 11:06:13.532143 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:13.797879 master-0 kubenswrapper[24928]: I1205 11:06:13.797799 24928 generic.go:334] "Generic (PLEG): container finished" podID="4ae72689-6505-4064-bd26-861bda2f68cc" containerID="349e2e75b98f5ec0c4eb43ce361e97dfbc1848d82e597f7be07ffc8be044fa44" exitCode=0 Dec 05 11:06:13.799641 master-0 kubenswrapper[24928]: I1205 11:06:13.799593 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-db-sync-9nbmc" event={"ID":"4ae72689-6505-4064-bd26-861bda2f68cc","Type":"ContainerDied","Data":"349e2e75b98f5ec0c4eb43ce361e97dfbc1848d82e597f7be07ffc8be044fa44"} Dec 05 11:06:13.799641 master-0 kubenswrapper[24928]: I1205 11:06:13.799643 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:06:13.800640 master-0 kubenswrapper[24928]: I1205 11:06:13.800606 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:13.800835 master-0 kubenswrapper[24928]: I1205 11:06:13.800802 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:06:13.947495 master-0 kubenswrapper[24928]: I1205 11:06:13.947226 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:14.419454 master-0 kubenswrapper[24928]: I1205 11:06:14.411520 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6dc4d774b6-b9nzv" podStartSLOduration=8.411500286 podStartE2EDuration="8.411500286s" podCreationTimestamp="2025-12-05 11:06:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:14.406470361 +0000 UTC m=+1134.409664212" watchObservedRunningTime="2025-12-05 11:06:14.411500286 +0000 UTC m=+1134.414694137" Dec 05 11:06:14.869814 master-0 kubenswrapper[24928]: I1205 11:06:14.867272 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-558cb7d6b-wgm2l"] Dec 05 11:06:15.236773 master-0 kubenswrapper[24928]: I1205 11:06:15.236658 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:06:15.521273 master-0 kubenswrapper[24928]: W1205 11:06:15.521194 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfddf0e6c_7977_4429_9015_732fd8dbb6a1.slice/crio-c42e7f420bef23b4105b5e3168b8c6a245f23a04964bd40194b0eb4e58b5bba3 WatchSource:0}: Error finding container c42e7f420bef23b4105b5e3168b8c6a245f23a04964bd40194b0eb4e58b5bba3: Status 404 returned error can't find the container with id c42e7f420bef23b4105b5e3168b8c6a245f23a04964bd40194b0eb4e58b5bba3 Dec 05 11:06:15.529770 master-0 kubenswrapper[24928]: I1205 11:06:15.527159 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-config-data\") pod \"4ae72689-6505-4064-bd26-861bda2f68cc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " Dec 05 11:06:15.529770 master-0 kubenswrapper[24928]: I1205 11:06:15.527241 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-db-sync-config-data\") pod \"4ae72689-6505-4064-bd26-861bda2f68cc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " Dec 05 11:06:15.529770 master-0 kubenswrapper[24928]: I1205 11:06:15.527280 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-scripts\") pod \"4ae72689-6505-4064-bd26-861bda2f68cc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " Dec 05 11:06:15.529770 master-0 kubenswrapper[24928]: I1205 11:06:15.527609 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-combined-ca-bundle\") pod \"4ae72689-6505-4064-bd26-861bda2f68cc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " Dec 05 11:06:15.529770 master-0 kubenswrapper[24928]: I1205 11:06:15.527644 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68qcn\" (UniqueName: \"kubernetes.io/projected/4ae72689-6505-4064-bd26-861bda2f68cc-kube-api-access-68qcn\") pod \"4ae72689-6505-4064-bd26-861bda2f68cc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " Dec 05 11:06:15.529770 master-0 kubenswrapper[24928]: I1205 11:06:15.527768 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ae72689-6505-4064-bd26-861bda2f68cc-etc-machine-id\") pod \"4ae72689-6505-4064-bd26-861bda2f68cc\" (UID: \"4ae72689-6505-4064-bd26-861bda2f68cc\") " Dec 05 11:06:15.539589 master-0 kubenswrapper[24928]: I1205 11:06:15.538617 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ae72689-6505-4064-bd26-861bda2f68cc-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4ae72689-6505-4064-bd26-861bda2f68cc" (UID: "4ae72689-6505-4064-bd26-861bda2f68cc"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:15.539589 master-0 kubenswrapper[24928]: I1205 11:06:15.539398 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ae72689-6505-4064-bd26-861bda2f68cc-kube-api-access-68qcn" (OuterVolumeSpecName: "kube-api-access-68qcn") pod "4ae72689-6505-4064-bd26-861bda2f68cc" (UID: "4ae72689-6505-4064-bd26-861bda2f68cc"). InnerVolumeSpecName "kube-api-access-68qcn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:15.549491 master-0 kubenswrapper[24928]: I1205 11:06:15.549244 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4ae72689-6505-4064-bd26-861bda2f68cc" (UID: "4ae72689-6505-4064-bd26-861bda2f68cc"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:15.549864 master-0 kubenswrapper[24928]: I1205 11:06:15.549833 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-scripts" (OuterVolumeSpecName: "scripts") pod "4ae72689-6505-4064-bd26-861bda2f68cc" (UID: "4ae72689-6505-4064-bd26-861bda2f68cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:15.549997 master-0 kubenswrapper[24928]: I1205 11:06:15.549968 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:06:15.589734 master-0 kubenswrapper[24928]: I1205 11:06:15.589685 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ae72689-6505-4064-bd26-861bda2f68cc" (UID: "4ae72689-6505-4064-bd26-861bda2f68cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:15.643456 master-0 kubenswrapper[24928]: I1205 11:06:15.631696 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:15.643890 master-0 kubenswrapper[24928]: I1205 11:06:15.643839 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68qcn\" (UniqueName: \"kubernetes.io/projected/4ae72689-6505-4064-bd26-861bda2f68cc-kube-api-access-68qcn\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:15.644000 master-0 kubenswrapper[24928]: I1205 11:06:15.643983 24928 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ae72689-6505-4064-bd26-861bda2f68cc-etc-machine-id\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:15.644106 master-0 kubenswrapper[24928]: I1205 11:06:15.644090 24928 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-db-sync-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:15.644208 master-0 kubenswrapper[24928]: I1205 11:06:15.644192 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:15.711454 master-0 kubenswrapper[24928]: I1205 11:06:15.709653 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-config-data" (OuterVolumeSpecName: "config-data") pod "4ae72689-6505-4064-bd26-861bda2f68cc" (UID: "4ae72689-6505-4064-bd26-861bda2f68cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:15.788764 master-0 kubenswrapper[24928]: I1205 11:06:15.787786 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ae72689-6505-4064-bd26-861bda2f68cc-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:15.855645 master-0 kubenswrapper[24928]: I1205 11:06:15.852064 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"fddf0e6c-7977-4429-9015-732fd8dbb6a1","Type":"ContainerStarted","Data":"c42e7f420bef23b4105b5e3168b8c6a245f23a04964bd40194b0eb4e58b5bba3"} Dec 05 11:06:15.855645 master-0 kubenswrapper[24928]: I1205 11:06:15.854784 24928 generic.go:334] "Generic (PLEG): container finished" podID="90315296-e6a2-4db3-aa3d-35af0d21a55e" containerID="d9d3a6f3db4e457acdf2eb32316c388679088a54bd3b8dee9f35a85b238f51c1" exitCode=0 Dec 05 11:06:15.855645 master-0 kubenswrapper[24928]: I1205 11:06:15.854893 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-djbqg" event={"ID":"90315296-e6a2-4db3-aa3d-35af0d21a55e","Type":"ContainerDied","Data":"d9d3a6f3db4e457acdf2eb32316c388679088a54bd3b8dee9f35a85b238f51c1"} Dec 05 11:06:15.857573 master-0 kubenswrapper[24928]: I1205 11:06:15.857085 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-558cb7d6b-wgm2l" event={"ID":"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5","Type":"ContainerStarted","Data":"fac789e1cc1a577ff594fd9a7846677f0d0eab429aac13172540371012a8ab11"} Dec 05 11:06:15.858401 master-0 kubenswrapper[24928]: I1205 11:06:15.858360 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:06:15.858401 master-0 kubenswrapper[24928]: I1205 11:06:15.858396 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:06:15.859575 master-0 kubenswrapper[24928]: I1205 11:06:15.859538 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-db-sync-9nbmc" Dec 05 11:06:15.859751 master-0 kubenswrapper[24928]: I1205 11:06:15.859667 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-db-sync-9nbmc" event={"ID":"4ae72689-6505-4064-bd26-861bda2f68cc","Type":"ContainerDied","Data":"612bf794def52a95657619b5c41bbfc0aa5e330095d97560da0f18adc3833336"} Dec 05 11:06:15.859805 master-0 kubenswrapper[24928]: I1205 11:06:15.859756 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="612bf794def52a95657619b5c41bbfc0aa5e330095d97560da0f18adc3833336" Dec 05 11:06:16.054246 master-0 kubenswrapper[24928]: I1205 11:06:16.054113 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:06:16.197882 master-0 kubenswrapper[24928]: I1205 11:06:16.197824 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:06:16.269614 master-0 kubenswrapper[24928]: I1205 11:06:16.265654 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ec941-scheduler-0"] Dec 05 11:06:16.269614 master-0 kubenswrapper[24928]: E1205 11:06:16.266375 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ae72689-6505-4064-bd26-861bda2f68cc" containerName="cinder-ec941-db-sync" Dec 05 11:06:16.269614 master-0 kubenswrapper[24928]: I1205 11:06:16.266392 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ae72689-6505-4064-bd26-861bda2f68cc" containerName="cinder-ec941-db-sync" Dec 05 11:06:16.269614 master-0 kubenswrapper[24928]: I1205 11:06:16.266615 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ae72689-6505-4064-bd26-861bda2f68cc" containerName="cinder-ec941-db-sync" Dec 05 11:06:16.269614 master-0 kubenswrapper[24928]: I1205 11:06:16.267809 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.277770 master-0 kubenswrapper[24928]: I1205 11:06:16.271623 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-scripts" Dec 05 11:06:16.277770 master-0 kubenswrapper[24928]: I1205 11:06:16.271849 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-config-data" Dec 05 11:06:16.277770 master-0 kubenswrapper[24928]: I1205 11:06:16.271954 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-scheduler-config-data" Dec 05 11:06:16.284549 master-0 kubenswrapper[24928]: I1205 11:06:16.284469 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-scheduler-0"] Dec 05 11:06:16.342557 master-0 kubenswrapper[24928]: I1205 11:06:16.327527 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ec941-volume-lvm-iscsi-0"] Dec 05 11:06:16.362464 master-0 kubenswrapper[24928]: I1205 11:06:16.357508 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.369446 master-0 kubenswrapper[24928]: I1205 11:06:16.364497 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-volume-lvm-iscsi-config-data" Dec 05 11:06:16.437195 master-0 kubenswrapper[24928]: I1205 11:06:16.426988 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c1e71ab-2601-4cf7-9423-dfb0d1469568-etc-machine-id\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.437195 master-0 kubenswrapper[24928]: I1205 11:06:16.432196 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-combined-ca-bundle\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.437195 master-0 kubenswrapper[24928]: I1205 11:06:16.432379 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data-custom\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.437195 master-0 kubenswrapper[24928]: I1205 11:06:16.432514 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lllmr\" (UniqueName: \"kubernetes.io/projected/6c1e71ab-2601-4cf7-9423-dfb0d1469568-kube-api-access-lllmr\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.437195 master-0 kubenswrapper[24928]: I1205 11:06:16.432735 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.437195 master-0 kubenswrapper[24928]: I1205 11:06:16.432860 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-scripts\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.437195 master-0 kubenswrapper[24928]: I1205 11:06:16.433276 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-volume-lvm-iscsi-0"] Dec 05 11:06:16.501446 master-0 kubenswrapper[24928]: I1205 11:06:16.496575 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-559ff6cd8f-5sc8w"] Dec 05 11:06:16.516399 master-0 kubenswrapper[24928]: I1205 11:06:16.513747 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.546956 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c1e71ab-2601-4cf7-9423-dfb0d1469568-etc-machine-id\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.547066 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-nvme\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.547132 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-run\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.547276 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-iscsi\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.547388 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c1e71ab-2601-4cf7-9423-dfb0d1469568-etc-machine-id\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.547556 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.549332 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-sys\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.549620 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-lib-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.549984 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-combined-ca-bundle\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.551643 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-559ff6cd8f-5sc8w"] Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.553589 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-machine-id\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.553678 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-lib-modules\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.553703 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-brick\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.553765 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data-custom\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.553810 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data-custom\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.553837 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-combined-ca-bundle\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.553860 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lllmr\" (UniqueName: \"kubernetes.io/projected/6c1e71ab-2601-4cf7-9423-dfb0d1469568-kube-api-access-lllmr\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.553896 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-scripts\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.553974 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-dev\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.554032 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.554094 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwtpm\" (UniqueName: \"kubernetes.io/projected/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-kube-api-access-bwtpm\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.554145 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-scripts\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.554199 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.557369 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-combined-ca-bundle\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.558737 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data-custom\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.558879 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-scripts\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.565456 master-0 kubenswrapper[24928]: I1205 11:06:16.561385 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.568277 master-0 kubenswrapper[24928]: I1205 11:06:16.568224 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ec941-backup-0"] Dec 05 11:06:16.578966 master-0 kubenswrapper[24928]: I1205 11:06:16.570861 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.594590 master-0 kubenswrapper[24928]: I1205 11:06:16.583844 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-backup-config-data" Dec 05 11:06:16.594590 master-0 kubenswrapper[24928]: I1205 11:06:16.588335 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lllmr\" (UniqueName: \"kubernetes.io/projected/6c1e71ab-2601-4cf7-9423-dfb0d1469568-kube-api-access-lllmr\") pod \"cinder-ec941-scheduler-0\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.657611 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-sb\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.657699 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-swift-storage-0\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.657764 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-nvme\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.657785 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-run\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.657827 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-iscsi\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.657856 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-config\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.657910 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.657932 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-sys\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.657949 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-svc\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.657982 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-lib-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.658083 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-machine-id\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.658116 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-lib-modules\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.658134 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-brick\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.658189 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data-custom\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.658214 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-combined-ca-bundle\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.658245 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-scripts\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.658301 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-dev\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.658325 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7c79\" (UniqueName: \"kubernetes.io/projected/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-kube-api-access-v7c79\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.658393 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwtpm\" (UniqueName: \"kubernetes.io/projected/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-kube-api-access-bwtpm\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.658459 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.658519 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-nb\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.659462 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-iscsi\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.661463 master-0 kubenswrapper[24928]: I1205 11:06:16.659525 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-lib-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.688462 master-0 kubenswrapper[24928]: I1205 11:06:16.688392 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-backup-0"] Dec 05 11:06:16.693471 master-0 kubenswrapper[24928]: I1205 11:06:16.692910 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-sys\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.693924 master-0 kubenswrapper[24928]: I1205 11:06:16.693858 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:16.694755 master-0 kubenswrapper[24928]: I1205 11:06:16.694718 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-run\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.694839 master-0 kubenswrapper[24928]: I1205 11:06:16.694780 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-machine-id\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.695031 master-0 kubenswrapper[24928]: I1205 11:06:16.694997 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-brick\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.695141 master-0 kubenswrapper[24928]: I1205 11:06:16.695099 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.695141 master-0 kubenswrapper[24928]: I1205 11:06:16.695121 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-dev\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.695297 master-0 kubenswrapper[24928]: I1205 11:06:16.695178 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-lib-modules\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.695700 master-0 kubenswrapper[24928]: I1205 11:06:16.695660 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-nvme\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.698559 master-0 kubenswrapper[24928]: I1205 11:06:16.698253 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-scripts\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.699148 master-0 kubenswrapper[24928]: I1205 11:06:16.699097 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data-custom\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.699731 master-0 kubenswrapper[24928]: I1205 11:06:16.699673 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-combined-ca-bundle\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.700460 master-0 kubenswrapper[24928]: I1205 11:06:16.700405 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.728971 master-0 kubenswrapper[24928]: I1205 11:06:16.728885 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwtpm\" (UniqueName: \"kubernetes.io/projected/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-kube-api-access-bwtpm\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:16.756094 master-0 kubenswrapper[24928]: I1205 11:06:16.756023 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ec941-api-0"] Dec 05 11:06:16.761850 master-0 kubenswrapper[24928]: I1205 11:06:16.761788 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-scripts\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762167 master-0 kubenswrapper[24928]: I1205 11:06:16.761861 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-iscsi\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762167 master-0 kubenswrapper[24928]: I1205 11:06:16.762070 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-lib-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762167 master-0 kubenswrapper[24928]: I1205 11:06:16.762118 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-nvme\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762167 master-0 kubenswrapper[24928]: I1205 11:06:16.762139 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5bkl\" (UniqueName: \"kubernetes.io/projected/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-kube-api-access-x5bkl\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762386 master-0 kubenswrapper[24928]: I1205 11:06:16.762266 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7c79\" (UniqueName: \"kubernetes.io/projected/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-kube-api-access-v7c79\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.762386 master-0 kubenswrapper[24928]: I1205 11:06:16.762317 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-machine-id\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762386 master-0 kubenswrapper[24928]: I1205 11:06:16.762350 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data-custom\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762562 master-0 kubenswrapper[24928]: I1205 11:06:16.762396 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-brick\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762562 master-0 kubenswrapper[24928]: I1205 11:06:16.762444 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-combined-ca-bundle\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762562 master-0 kubenswrapper[24928]: I1205 11:06:16.762521 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-nb\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.762562 master-0 kubenswrapper[24928]: I1205 11:06:16.762550 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-sys\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762861 master-0 kubenswrapper[24928]: I1205 11:06:16.762579 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762861 master-0 kubenswrapper[24928]: I1205 11:06:16.762623 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-sb\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.762861 master-0 kubenswrapper[24928]: I1205 11:06:16.762658 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-swift-storage-0\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.762861 master-0 kubenswrapper[24928]: I1205 11:06:16.762713 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-run\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762861 master-0 kubenswrapper[24928]: I1205 11:06:16.762742 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-dev\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762861 master-0 kubenswrapper[24928]: I1205 11:06:16.762769 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-lib-modules\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.762861 master-0 kubenswrapper[24928]: I1205 11:06:16.762803 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-config\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.762861 master-0 kubenswrapper[24928]: I1205 11:06:16.762830 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.763347 master-0 kubenswrapper[24928]: I1205 11:06:16.762868 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-svc\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.764391 master-0 kubenswrapper[24928]: I1205 11:06:16.764352 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-svc\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.766458 master-0 kubenswrapper[24928]: I1205 11:06:16.765964 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-nb\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.766649 master-0 kubenswrapper[24928]: I1205 11:06:16.766465 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-config\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.766649 master-0 kubenswrapper[24928]: I1205 11:06:16.766545 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-sb\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.767443 master-0 kubenswrapper[24928]: I1205 11:06:16.767065 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-swift-storage-0\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.774285 master-0 kubenswrapper[24928]: I1205 11:06:16.774224 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-api-0"] Dec 05 11:06:16.774644 master-0 kubenswrapper[24928]: I1205 11:06:16.774353 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-api-0" Dec 05 11:06:16.777373 master-0 kubenswrapper[24928]: I1205 11:06:16.776878 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-api-config-data" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.892793 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-nvme\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.892869 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5bkl\" (UniqueName: \"kubernetes.io/projected/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-kube-api-access-x5bkl\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.892935 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9675e13-fa09-4eac-873d-5b8068371c3f-logs\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.892966 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28ssg\" (UniqueName: \"kubernetes.io/projected/a9675e13-fa09-4eac-873d-5b8068371c3f-kube-api-access-28ssg\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893033 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-combined-ca-bundle\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893083 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-machine-id\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893117 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data-custom\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893145 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-brick\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893177 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-combined-ca-bundle\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893250 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-sys\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893274 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893303 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-scripts\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893331 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data-custom\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893373 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a9675e13-fa09-4eac-873d-5b8068371c3f-etc-machine-id\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893435 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-run\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893459 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893486 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-dev\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893510 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-lib-modules\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.893908 master-0 kubenswrapper[24928]: I1205 11:06:16.893536 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.895494 master-0 kubenswrapper[24928]: I1205 11:06:16.895137 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-scripts\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.895494 master-0 kubenswrapper[24928]: I1205 11:06:16.895193 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-iscsi\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.895494 master-0 kubenswrapper[24928]: I1205 11:06:16.895232 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-lib-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.895494 master-0 kubenswrapper[24928]: I1205 11:06:16.895436 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-lib-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.895717 master-0 kubenswrapper[24928]: I1205 11:06:16.895537 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-run\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.895717 master-0 kubenswrapper[24928]: I1205 11:06:16.895672 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-dev\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.897909 master-0 kubenswrapper[24928]: I1205 11:06:16.895973 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-brick\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.897909 master-0 kubenswrapper[24928]: I1205 11:06:16.896043 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.897909 master-0 kubenswrapper[24928]: I1205 11:06:16.896574 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-lib-modules\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.897909 master-0 kubenswrapper[24928]: I1205 11:06:16.896589 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-nvme\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.897909 master-0 kubenswrapper[24928]: I1205 11:06:16.896644 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-sys\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.897909 master-0 kubenswrapper[24928]: I1205 11:06:16.896677 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-machine-id\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.897909 master-0 kubenswrapper[24928]: I1205 11:06:16.896805 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-iscsi\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.908350 master-0 kubenswrapper[24928]: I1205 11:06:16.902656 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-scripts\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.912563 master-0 kubenswrapper[24928]: I1205 11:06:16.910888 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.913346 master-0 kubenswrapper[24928]: I1205 11:06:16.913275 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data-custom\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.917353 master-0 kubenswrapper[24928]: I1205 11:06:16.917295 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-combined-ca-bundle\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.923237 master-0 kubenswrapper[24928]: I1205 11:06:16.922811 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7c79\" (UniqueName: \"kubernetes.io/projected/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-kube-api-access-v7c79\") pod \"dnsmasq-dns-559ff6cd8f-5sc8w\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.926312 master-0 kubenswrapper[24928]: I1205 11:06:16.926259 24928 generic.go:334] "Generic (PLEG): container finished" podID="0e18fdd2-be67-4228-82a3-4d02d3b350cf" containerID="d3b829955ca0a2f950c80b244ce91bc30a45fc6ecffaf9ceea3e6d188ed8be3c" exitCode=0 Dec 05 11:06:16.928864 master-0 kubenswrapper[24928]: I1205 11:06:16.926346 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-sxns9" event={"ID":"0e18fdd2-be67-4228-82a3-4d02d3b350cf","Type":"ContainerDied","Data":"d3b829955ca0a2f950c80b244ce91bc30a45fc6ecffaf9ceea3e6d188ed8be3c"} Dec 05 11:06:16.928864 master-0 kubenswrapper[24928]: I1205 11:06:16.927795 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5bkl\" (UniqueName: \"kubernetes.io/projected/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-kube-api-access-x5bkl\") pod \"cinder-ec941-backup-0\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:16.977145 master-0 kubenswrapper[24928]: I1205 11:06:16.965994 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:16.980405 master-0 kubenswrapper[24928]: I1205 11:06:16.979513 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-558cb7d6b-wgm2l" event={"ID":"1a7dd686-7e28-44e0-9cd9-7ac8aa456cf5","Type":"ContainerStarted","Data":"0a2096b496d1978ee9a5e4876f3f1b2b6792414517b03dd65115d74b405d6dfc"} Dec 05 11:06:16.980876 master-0 kubenswrapper[24928]: I1205 11:06:16.980807 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:16.990957 master-0 kubenswrapper[24928]: I1205 11:06:16.988836 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:17.000452 master-0 kubenswrapper[24928]: I1205 11:06:16.997351 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9675e13-fa09-4eac-873d-5b8068371c3f-logs\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.000452 master-0 kubenswrapper[24928]: I1205 11:06:16.997467 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28ssg\" (UniqueName: \"kubernetes.io/projected/a9675e13-fa09-4eac-873d-5b8068371c3f-kube-api-access-28ssg\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.000452 master-0 kubenswrapper[24928]: I1205 11:06:16.997533 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-combined-ca-bundle\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.000452 master-0 kubenswrapper[24928]: I1205 11:06:16.997621 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-scripts\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.000452 master-0 kubenswrapper[24928]: I1205 11:06:16.997644 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data-custom\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.000452 master-0 kubenswrapper[24928]: I1205 11:06:16.997670 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a9675e13-fa09-4eac-873d-5b8068371c3f-etc-machine-id\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.000452 master-0 kubenswrapper[24928]: I1205 11:06:16.997698 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.000963 master-0 kubenswrapper[24928]: I1205 11:06:17.000673 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9675e13-fa09-4eac-873d-5b8068371c3f-logs\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.003794 master-0 kubenswrapper[24928]: I1205 11:06:17.003071 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a9675e13-fa09-4eac-873d-5b8068371c3f-etc-machine-id\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.007893 master-0 kubenswrapper[24928]: I1205 11:06:17.007079 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-558cb7d6b-wgm2l" podStartSLOduration=7.007045772 podStartE2EDuration="7.007045772s" podCreationTimestamp="2025-12-05 11:06:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:17.001877522 +0000 UTC m=+1137.005071383" watchObservedRunningTime="2025-12-05 11:06:17.007045772 +0000 UTC m=+1137.010239613" Dec 05 11:06:17.016883 master-0 kubenswrapper[24928]: I1205 11:06:17.016805 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.020456 master-0 kubenswrapper[24928]: I1205 11:06:17.019132 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-scripts\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.020456 master-0 kubenswrapper[24928]: I1205 11:06:17.019385 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-combined-ca-bundle\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.020456 master-0 kubenswrapper[24928]: I1205 11:06:17.020245 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:17.023034 master-0 kubenswrapper[24928]: I1205 11:06:17.022238 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data-custom\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.030875 master-0 kubenswrapper[24928]: I1205 11:06:17.030800 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28ssg\" (UniqueName: \"kubernetes.io/projected/a9675e13-fa09-4eac-873d-5b8068371c3f-kube-api-access-28ssg\") pod \"cinder-ec941-api-0\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.233817 master-0 kubenswrapper[24928]: I1205 11:06:17.233385 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-api-0" Dec 05 11:06:17.308244 master-0 kubenswrapper[24928]: I1205 11:06:17.308151 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-scheduler-0"] Dec 05 11:06:17.389401 master-0 kubenswrapper[24928]: W1205 11:06:17.389351 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c1e71ab_2601_4cf7_9423_dfb0d1469568.slice/crio-137e4aab0a747fe27cee7d03d27e72129c68fc8b5e6cf6ebe66d2947fcbbf24a WatchSource:0}: Error finding container 137e4aab0a747fe27cee7d03d27e72129c68fc8b5e6cf6ebe66d2947fcbbf24a: Status 404 returned error can't find the container with id 137e4aab0a747fe27cee7d03d27e72129c68fc8b5e6cf6ebe66d2947fcbbf24a Dec 05 11:06:17.550906 master-0 kubenswrapper[24928]: I1205 11:06:17.550864 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-djbqg" Dec 05 11:06:17.718827 master-0 kubenswrapper[24928]: I1205 11:06:17.718680 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fdzk\" (UniqueName: \"kubernetes.io/projected/90315296-e6a2-4db3-aa3d-35af0d21a55e-kube-api-access-7fdzk\") pod \"90315296-e6a2-4db3-aa3d-35af0d21a55e\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " Dec 05 11:06:17.718994 master-0 kubenswrapper[24928]: I1205 11:06:17.718850 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-config\") pod \"90315296-e6a2-4db3-aa3d-35af0d21a55e\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " Dec 05 11:06:17.719086 master-0 kubenswrapper[24928]: I1205 11:06:17.719059 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-combined-ca-bundle\") pod \"90315296-e6a2-4db3-aa3d-35af0d21a55e\" (UID: \"90315296-e6a2-4db3-aa3d-35af0d21a55e\") " Dec 05 11:06:17.726036 master-0 kubenswrapper[24928]: I1205 11:06:17.725963 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90315296-e6a2-4db3-aa3d-35af0d21a55e-kube-api-access-7fdzk" (OuterVolumeSpecName: "kube-api-access-7fdzk") pod "90315296-e6a2-4db3-aa3d-35af0d21a55e" (UID: "90315296-e6a2-4db3-aa3d-35af0d21a55e"). InnerVolumeSpecName "kube-api-access-7fdzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:17.766747 master-0 kubenswrapper[24928]: I1205 11:06:17.766185 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90315296-e6a2-4db3-aa3d-35af0d21a55e" (UID: "90315296-e6a2-4db3-aa3d-35af0d21a55e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:17.772826 master-0 kubenswrapper[24928]: I1205 11:06:17.772744 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-config" (OuterVolumeSpecName: "config") pod "90315296-e6a2-4db3-aa3d-35af0d21a55e" (UID: "90315296-e6a2-4db3-aa3d-35af0d21a55e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:17.823191 master-0 kubenswrapper[24928]: I1205 11:06:17.822126 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:17.823191 master-0 kubenswrapper[24928]: I1205 11:06:17.822191 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90315296-e6a2-4db3-aa3d-35af0d21a55e-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:17.823191 master-0 kubenswrapper[24928]: I1205 11:06:17.822209 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fdzk\" (UniqueName: \"kubernetes.io/projected/90315296-e6a2-4db3-aa3d-35af0d21a55e-kube-api-access-7fdzk\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:17.908349 master-0 kubenswrapper[24928]: I1205 11:06:17.906102 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-volume-lvm-iscsi-0"] Dec 05 11:06:17.910705 master-0 kubenswrapper[24928]: W1205 11:06:17.910627 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf499b5c5_b9e1_4609_8dd1_8ce88a97c77b.slice/crio-2a83e06b47b03e0809f8cdbfcd31100606a90720161d5407ad1311a27cd13656 WatchSource:0}: Error finding container 2a83e06b47b03e0809f8cdbfcd31100606a90720161d5407ad1311a27cd13656: Status 404 returned error can't find the container with id 2a83e06b47b03e0809f8cdbfcd31100606a90720161d5407ad1311a27cd13656 Dec 05 11:06:17.920082 master-0 kubenswrapper[24928]: I1205 11:06:17.920015 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-559ff6cd8f-5sc8w"] Dec 05 11:06:18.006766 master-0 kubenswrapper[24928]: I1205 11:06:18.006689 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" event={"ID":"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3","Type":"ContainerStarted","Data":"919395ea67fe52035f25c7eb3223cd04ba25cf797eb7c4ca6dbb35003545ae21"} Dec 05 11:06:18.008430 master-0 kubenswrapper[24928]: I1205 11:06:18.008362 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" event={"ID":"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b","Type":"ContainerStarted","Data":"2a83e06b47b03e0809f8cdbfcd31100606a90720161d5407ad1311a27cd13656"} Dec 05 11:06:18.029540 master-0 kubenswrapper[24928]: I1205 11:06:18.029155 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-sxns9" event={"ID":"0e18fdd2-be67-4228-82a3-4d02d3b350cf","Type":"ContainerStarted","Data":"a33daad2110904d887bdef13f8dcdbc0e41b61ab952bb4b352f57727fbcd6e6e"} Dec 05 11:06:18.040333 master-0 kubenswrapper[24928]: I1205 11:06:18.040261 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"fddf0e6c-7977-4429-9015-732fd8dbb6a1","Type":"ContainerStarted","Data":"7c116d246b2d3f460a0aa28456f2c1fe5c53a3c50f6ed7a5a3f758d1ad2cb7c5"} Dec 05 11:06:18.070155 master-0 kubenswrapper[24928]: I1205 11:06:18.070096 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-djbqg" Dec 05 11:06:18.070476 master-0 kubenswrapper[24928]: I1205 11:06:18.070438 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-djbqg" event={"ID":"90315296-e6a2-4db3-aa3d-35af0d21a55e","Type":"ContainerDied","Data":"003031ae2b91b805735e82c853f5e2c92779ac6eca6d6175ce5ca5793a689b4f"} Dec 05 11:06:18.070546 master-0 kubenswrapper[24928]: I1205 11:06:18.070507 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="003031ae2b91b805735e82c853f5e2c92779ac6eca6d6175ce5ca5793a689b4f" Dec 05 11:06:18.099969 master-0 kubenswrapper[24928]: I1205 11:06:18.099870 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-scheduler-0" event={"ID":"6c1e71ab-2601-4cf7-9423-dfb0d1469568","Type":"ContainerStarted","Data":"137e4aab0a747fe27cee7d03d27e72129c68fc8b5e6cf6ebe66d2947fcbbf24a"} Dec 05 11:06:18.123189 master-0 kubenswrapper[24928]: I1205 11:06:18.123005 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-db-sync-sxns9" podStartSLOduration=16.947163008 podStartE2EDuration="28.122981654s" podCreationTimestamp="2025-12-05 11:05:50 +0000 UTC" firstStartedPulling="2025-12-05 11:06:04.295913663 +0000 UTC m=+1124.299107514" lastFinishedPulling="2025-12-05 11:06:15.471732289 +0000 UTC m=+1135.474926160" observedRunningTime="2025-12-05 11:06:18.082871773 +0000 UTC m=+1138.086065644" watchObservedRunningTime="2025-12-05 11:06:18.122981654 +0000 UTC m=+1138.126175505" Dec 05 11:06:18.151778 master-0 kubenswrapper[24928]: I1205 11:06:18.149342 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-backup-0"] Dec 05 11:06:18.182451 master-0 kubenswrapper[24928]: I1205 11:06:18.180700 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-api-0"] Dec 05 11:06:18.229098 master-0 kubenswrapper[24928]: I1205 11:06:18.196912 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-559ff6cd8f-5sc8w"] Dec 05 11:06:18.296826 master-0 kubenswrapper[24928]: I1205 11:06:18.296761 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59b84c6dcc-54hsm"] Dec 05 11:06:18.297305 master-0 kubenswrapper[24928]: E1205 11:06:18.297203 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90315296-e6a2-4db3-aa3d-35af0d21a55e" containerName="neutron-db-sync" Dec 05 11:06:18.297305 master-0 kubenswrapper[24928]: I1205 11:06:18.297246 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="90315296-e6a2-4db3-aa3d-35af0d21a55e" containerName="neutron-db-sync" Dec 05 11:06:18.297610 master-0 kubenswrapper[24928]: I1205 11:06:18.297579 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="90315296-e6a2-4db3-aa3d-35af0d21a55e" containerName="neutron-db-sync" Dec 05 11:06:18.299192 master-0 kubenswrapper[24928]: I1205 11:06:18.299154 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59b84c6dcc-54hsm"] Dec 05 11:06:18.299358 master-0 kubenswrapper[24928]: I1205 11:06:18.299297 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.431835 master-0 kubenswrapper[24928]: I1205 11:06:18.431750 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7bd67d49b4-fmw2f"] Dec 05 11:06:18.435220 master-0 kubenswrapper[24928]: I1205 11:06:18.435159 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.438005 master-0 kubenswrapper[24928]: I1205 11:06:18.437950 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 11:06:18.438255 master-0 kubenswrapper[24928]: I1205 11:06:18.438211 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 05 11:06:18.438497 master-0 kubenswrapper[24928]: I1205 11:06:18.438476 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 11:06:18.518665 master-0 kubenswrapper[24928]: I1205 11:06:18.518403 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-combined-ca-bundle\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.518665 master-0 kubenswrapper[24928]: I1205 11:06:18.518554 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-httpd-config\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.518665 master-0 kubenswrapper[24928]: I1205 11:06:18.518597 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-ovndb-tls-certs\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.518665 master-0 kubenswrapper[24928]: I1205 11:06:18.518618 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-nb\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.518961 master-0 kubenswrapper[24928]: I1205 11:06:18.518831 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-swift-storage-0\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.518961 master-0 kubenswrapper[24928]: I1205 11:06:18.518894 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-config\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.518961 master-0 kubenswrapper[24928]: I1205 11:06:18.518938 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfkp5\" (UniqueName: \"kubernetes.io/projected/81360b3f-3235-4cc7-9ef3-e8f748037762-kube-api-access-nfkp5\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.519065 master-0 kubenswrapper[24928]: I1205 11:06:18.519007 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpmht\" (UniqueName: \"kubernetes.io/projected/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-kube-api-access-gpmht\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.519117 master-0 kubenswrapper[24928]: I1205 11:06:18.519083 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-config\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.519179 master-0 kubenswrapper[24928]: I1205 11:06:18.519146 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-sb\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.519179 master-0 kubenswrapper[24928]: I1205 11:06:18.519166 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-svc\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.538808 master-0 kubenswrapper[24928]: I1205 11:06:18.536183 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bd67d49b4-fmw2f"] Dec 05 11:06:18.632794 master-0 kubenswrapper[24928]: I1205 11:06:18.630630 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-httpd-config\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.632794 master-0 kubenswrapper[24928]: I1205 11:06:18.630747 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-ovndb-tls-certs\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.632794 master-0 kubenswrapper[24928]: I1205 11:06:18.630784 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-nb\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.632794 master-0 kubenswrapper[24928]: I1205 11:06:18.630907 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-swift-storage-0\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.632794 master-0 kubenswrapper[24928]: I1205 11:06:18.630935 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-config\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.632794 master-0 kubenswrapper[24928]: I1205 11:06:18.630966 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfkp5\" (UniqueName: \"kubernetes.io/projected/81360b3f-3235-4cc7-9ef3-e8f748037762-kube-api-access-nfkp5\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.632794 master-0 kubenswrapper[24928]: I1205 11:06:18.631007 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpmht\" (UniqueName: \"kubernetes.io/projected/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-kube-api-access-gpmht\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.632794 master-0 kubenswrapper[24928]: I1205 11:06:18.631044 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-config\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.632794 master-0 kubenswrapper[24928]: I1205 11:06:18.631130 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-sb\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.632794 master-0 kubenswrapper[24928]: I1205 11:06:18.631150 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-svc\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.632794 master-0 kubenswrapper[24928]: I1205 11:06:18.631200 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-combined-ca-bundle\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.637489 master-0 kubenswrapper[24928]: I1205 11:06:18.637182 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-httpd-config\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.662985 master-0 kubenswrapper[24928]: I1205 11:06:18.642174 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-config\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.662985 master-0 kubenswrapper[24928]: I1205 11:06:18.658849 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfkp5\" (UniqueName: \"kubernetes.io/projected/81360b3f-3235-4cc7-9ef3-e8f748037762-kube-api-access-nfkp5\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.662985 master-0 kubenswrapper[24928]: I1205 11:06:18.658961 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-ovndb-tls-certs\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.662985 master-0 kubenswrapper[24928]: I1205 11:06:18.661668 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-config\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.662985 master-0 kubenswrapper[24928]: I1205 11:06:18.662160 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpmht\" (UniqueName: \"kubernetes.io/projected/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-kube-api-access-gpmht\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.687753 master-0 kubenswrapper[24928]: I1205 11:06:18.667191 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-nb\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.687753 master-0 kubenswrapper[24928]: I1205 11:06:18.670549 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-combined-ca-bundle\") pod \"neutron-7bd67d49b4-fmw2f\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.687753 master-0 kubenswrapper[24928]: I1205 11:06:18.687339 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-sb\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.687753 master-0 kubenswrapper[24928]: I1205 11:06:18.687387 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-swift-storage-0\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.710484 master-0 kubenswrapper[24928]: I1205 11:06:18.688301 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:18.710484 master-0 kubenswrapper[24928]: I1205 11:06:18.693262 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-svc\") pod \"dnsmasq-dns-59b84c6dcc-54hsm\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:18.919770 master-0 kubenswrapper[24928]: I1205 11:06:18.919707 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:19.126449 master-0 kubenswrapper[24928]: I1205 11:06:19.126380 24928 generic.go:334] "Generic (PLEG): container finished" podID="a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" containerID="8e41195db08db19ac9aab7613e4bbebfd1d136cad07fd144998fb4cafbffe797" exitCode=0 Dec 05 11:06:19.126648 master-0 kubenswrapper[24928]: I1205 11:06:19.126488 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" event={"ID":"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3","Type":"ContainerDied","Data":"8e41195db08db19ac9aab7613e4bbebfd1d136cad07fd144998fb4cafbffe797"} Dec 05 11:06:19.134464 master-0 kubenswrapper[24928]: I1205 11:06:19.134384 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-api-0" event={"ID":"a9675e13-fa09-4eac-873d-5b8068371c3f","Type":"ContainerStarted","Data":"be5a6630c19ef5dd4a1814f19f09a0c93e861d0d9a7a6663e450d230d9055421"} Dec 05 11:06:19.140720 master-0 kubenswrapper[24928]: I1205 11:06:19.140642 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"fddf0e6c-7977-4429-9015-732fd8dbb6a1","Type":"ContainerStarted","Data":"2130b735befea54638f55b346f56e9884c2c0ab93e77888bb7dad2d68c56bea0"} Dec 05 11:06:19.146211 master-0 kubenswrapper[24928]: I1205 11:06:19.146148 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-backup-0" event={"ID":"9e8aa5cf-e430-4c19-8eca-fe9269299b5b","Type":"ContainerStarted","Data":"3598120c56fe1399261a2d8197455576335b521ed3a81be8916ef237494a733a"} Dec 05 11:06:19.601141 master-0 kubenswrapper[24928]: I1205 11:06:19.600946 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-ec941-default-internal-api-0" podStartSLOduration=14.600925916 podStartE2EDuration="14.600925916s" podCreationTimestamp="2025-12-05 11:06:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:19.192004777 +0000 UTC m=+1139.195198638" watchObservedRunningTime="2025-12-05 11:06:19.600925916 +0000 UTC m=+1139.604119767" Dec 05 11:06:19.615594 master-0 kubenswrapper[24928]: I1205 11:06:19.611473 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ec941-api-0"] Dec 05 11:06:19.774352 master-0 kubenswrapper[24928]: I1205 11:06:19.765685 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:19.910077 master-0 kubenswrapper[24928]: I1205 11:06:19.910018 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-swift-storage-0\") pod \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " Dec 05 11:06:19.910364 master-0 kubenswrapper[24928]: I1205 11:06:19.910332 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-nb\") pod \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " Dec 05 11:06:19.910524 master-0 kubenswrapper[24928]: I1205 11:06:19.910459 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-svc\") pod \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " Dec 05 11:06:19.910524 master-0 kubenswrapper[24928]: I1205 11:06:19.910496 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-config\") pod \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " Dec 05 11:06:19.910650 master-0 kubenswrapper[24928]: I1205 11:06:19.910540 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-sb\") pod \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " Dec 05 11:06:19.910650 master-0 kubenswrapper[24928]: I1205 11:06:19.910601 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7c79\" (UniqueName: \"kubernetes.io/projected/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-kube-api-access-v7c79\") pod \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\" (UID: \"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3\") " Dec 05 11:06:19.919767 master-0 kubenswrapper[24928]: I1205 11:06:19.919708 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-kube-api-access-v7c79" (OuterVolumeSpecName: "kube-api-access-v7c79") pod "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" (UID: "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3"). InnerVolumeSpecName "kube-api-access-v7c79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:19.950185 master-0 kubenswrapper[24928]: I1205 11:06:19.950136 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" (UID: "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:19.960530 master-0 kubenswrapper[24928]: I1205 11:06:19.960405 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" (UID: "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:19.974056 master-0 kubenswrapper[24928]: I1205 11:06:19.973785 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" (UID: "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:19.987949 master-0 kubenswrapper[24928]: I1205 11:06:19.974589 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-config" (OuterVolumeSpecName: "config") pod "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" (UID: "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:19.989075 master-0 kubenswrapper[24928]: I1205 11:06:19.988676 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" (UID: "a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:19.997347 master-0 kubenswrapper[24928]: I1205 11:06:19.994866 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59b84c6dcc-54hsm"] Dec 05 11:06:20.015315 master-0 kubenswrapper[24928]: I1205 11:06:20.014578 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-sb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:20.015315 master-0 kubenswrapper[24928]: I1205 11:06:20.014615 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7c79\" (UniqueName: \"kubernetes.io/projected/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-kube-api-access-v7c79\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:20.015315 master-0 kubenswrapper[24928]: I1205 11:06:20.014627 24928 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-swift-storage-0\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:20.015315 master-0 kubenswrapper[24928]: I1205 11:06:20.014636 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:20.015315 master-0 kubenswrapper[24928]: I1205 11:06:20.014645 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:20.015315 master-0 kubenswrapper[24928]: I1205 11:06:20.014654 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:20.035457 master-0 kubenswrapper[24928]: W1205 11:06:20.035267 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc684b0cd_620b_45f3_9fee_4eb6f8cf99df.slice/crio-c2d9e6b2ccef35e52011744ae9a644a93ff1041922f84f34e5e7840bd1d9afa2 WatchSource:0}: Error finding container c2d9e6b2ccef35e52011744ae9a644a93ff1041922f84f34e5e7840bd1d9afa2: Status 404 returned error can't find the container with id c2d9e6b2ccef35e52011744ae9a644a93ff1041922f84f34e5e7840bd1d9afa2 Dec 05 11:06:20.160821 master-0 kubenswrapper[24928]: I1205 11:06:20.160759 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" event={"ID":"c684b0cd-620b-45f3-9fee-4eb6f8cf99df","Type":"ContainerStarted","Data":"c2d9e6b2ccef35e52011744ae9a644a93ff1041922f84f34e5e7840bd1d9afa2"} Dec 05 11:06:20.162724 master-0 kubenswrapper[24928]: I1205 11:06:20.162693 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-backup-0" event={"ID":"9e8aa5cf-e430-4c19-8eca-fe9269299b5b","Type":"ContainerStarted","Data":"f26dac4d5efacba623abb0a2e976dc18d512c8b3e96f5c357dd52cdef8fa5976"} Dec 05 11:06:20.164901 master-0 kubenswrapper[24928]: I1205 11:06:20.164864 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-scheduler-0" event={"ID":"6c1e71ab-2601-4cf7-9423-dfb0d1469568","Type":"ContainerStarted","Data":"9ea11651e662d55a392a4b5060336712eeda3be70b16b2e1d9fa94328693f8ab"} Dec 05 11:06:20.171069 master-0 kubenswrapper[24928]: I1205 11:06:20.171024 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:20.171201 master-0 kubenswrapper[24928]: I1205 11:06:20.171053 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" event={"ID":"a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3","Type":"ContainerDied","Data":"919395ea67fe52035f25c7eb3223cd04ba25cf797eb7c4ca6dbb35003545ae21"} Dec 05 11:06:20.171201 master-0 kubenswrapper[24928]: I1205 11:06:20.171123 24928 scope.go:117] "RemoveContainer" containerID="8e41195db08db19ac9aab7613e4bbebfd1d136cad07fd144998fb4cafbffe797" Dec 05 11:06:20.176733 master-0 kubenswrapper[24928]: I1205 11:06:20.176676 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" event={"ID":"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b","Type":"ContainerStarted","Data":"489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95"} Dec 05 11:06:20.180595 master-0 kubenswrapper[24928]: I1205 11:06:20.180530 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-api-0" event={"ID":"a9675e13-fa09-4eac-873d-5b8068371c3f","Type":"ContainerStarted","Data":"aac6390a98833664c6898788e300db78fe0757cf92c0a297830d83be031a821a"} Dec 05 11:06:21.196083 master-0 kubenswrapper[24928]: I1205 11:06:21.196019 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" event={"ID":"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b","Type":"ContainerStarted","Data":"2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a"} Dec 05 11:06:21.535645 master-0 kubenswrapper[24928]: I1205 11:06:21.535367 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7bd67d49b4-fmw2f"] Dec 05 11:06:21.557466 master-0 kubenswrapper[24928]: W1205 11:06:21.556157 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81360b3f_3235_4cc7_9ef3_e8f748037762.slice/crio-9db0491738dd6eeec3d1bb79bd094595b7517436950f2ab2ff8e9181b13fba21 WatchSource:0}: Error finding container 9db0491738dd6eeec3d1bb79bd094595b7517436950f2ab2ff8e9181b13fba21: Status 404 returned error can't find the container with id 9db0491738dd6eeec3d1bb79bd094595b7517436950f2ab2ff8e9181b13fba21 Dec 05 11:06:22.220128 master-0 kubenswrapper[24928]: I1205 11:06:22.219998 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-ec941-api-0" podUID="a9675e13-fa09-4eac-873d-5b8068371c3f" containerName="cinder-ec941-api-log" containerID="cri-o://aac6390a98833664c6898788e300db78fe0757cf92c0a297830d83be031a821a" gracePeriod=30 Dec 05 11:06:22.220786 master-0 kubenswrapper[24928]: I1205 11:06:22.220670 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-ec941-api-0" podUID="a9675e13-fa09-4eac-873d-5b8068371c3f" containerName="cinder-api" containerID="cri-o://9149da194cd2fcace386cbee97d69b5b55685679f29ab434fe5ce256a9f87df1" gracePeriod=30 Dec 05 11:06:22.220786 master-0 kubenswrapper[24928]: I1205 11:06:22.220753 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-ec941-api-0" Dec 05 11:06:22.220896 master-0 kubenswrapper[24928]: I1205 11:06:22.220801 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-api-0" event={"ID":"a9675e13-fa09-4eac-873d-5b8068371c3f","Type":"ContainerStarted","Data":"9149da194cd2fcace386cbee97d69b5b55685679f29ab434fe5ce256a9f87df1"} Dec 05 11:06:22.224462 master-0 kubenswrapper[24928]: I1205 11:06:22.223531 24928 generic.go:334] "Generic (PLEG): container finished" podID="c684b0cd-620b-45f3-9fee-4eb6f8cf99df" containerID="b259bd55744f87238af7ef1659a02b7493e7a2fd6bf68f40b00bbea8b402739a" exitCode=0 Dec 05 11:06:22.224462 master-0 kubenswrapper[24928]: I1205 11:06:22.223613 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" event={"ID":"c684b0cd-620b-45f3-9fee-4eb6f8cf99df","Type":"ContainerDied","Data":"b259bd55744f87238af7ef1659a02b7493e7a2fd6bf68f40b00bbea8b402739a"} Dec 05 11:06:22.237699 master-0 kubenswrapper[24928]: I1205 11:06:22.237603 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd67d49b4-fmw2f" event={"ID":"81360b3f-3235-4cc7-9ef3-e8f748037762","Type":"ContainerStarted","Data":"e086a3c64a2da4fb8778731203db97137f5bdf9bf355be905b73134e8fd137d3"} Dec 05 11:06:22.237699 master-0 kubenswrapper[24928]: I1205 11:06:22.237665 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd67d49b4-fmw2f" event={"ID":"81360b3f-3235-4cc7-9ef3-e8f748037762","Type":"ContainerStarted","Data":"66554b0348b461b3687234b1ec9f9762acaa91069b3301f211e96317a7da9a8e"} Dec 05 11:06:22.237699 master-0 kubenswrapper[24928]: I1205 11:06:22.237675 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd67d49b4-fmw2f" event={"ID":"81360b3f-3235-4cc7-9ef3-e8f748037762","Type":"ContainerStarted","Data":"9db0491738dd6eeec3d1bb79bd094595b7517436950f2ab2ff8e9181b13fba21"} Dec 05 11:06:22.238722 master-0 kubenswrapper[24928]: I1205 11:06:22.238666 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:22.243667 master-0 kubenswrapper[24928]: I1205 11:06:22.243525 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-backup-0" event={"ID":"9e8aa5cf-e430-4c19-8eca-fe9269299b5b","Type":"ContainerStarted","Data":"b0bfa1ee2a8e48fd72a6d0f74a59625ab9f56c7d9d3527d121ba22124b79d94e"} Dec 05 11:06:22.249525 master-0 kubenswrapper[24928]: I1205 11:06:22.249255 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-scheduler-0" event={"ID":"6c1e71ab-2601-4cf7-9423-dfb0d1469568","Type":"ContainerStarted","Data":"8acdbec23abbc42cb9d573312a4c8ab4f72724ac083ac0a0d135ed35c957154b"} Dec 05 11:06:22.597127 master-0 kubenswrapper[24928]: I1205 11:06:22.597017 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-ec941-api-0" podStartSLOduration=6.59699526 podStartE2EDuration="6.59699526s" podCreationTimestamp="2025-12-05 11:06:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:22.575763351 +0000 UTC m=+1142.578957202" watchObservedRunningTime="2025-12-05 11:06:22.59699526 +0000 UTC m=+1142.600189121" Dec 05 11:06:22.667105 master-0 kubenswrapper[24928]: I1205 11:06:22.667019 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-ec941-scheduler-0" podStartSLOduration=5.52691599 podStartE2EDuration="6.667001386s" podCreationTimestamp="2025-12-05 11:06:16 +0000 UTC" firstStartedPulling="2025-12-05 11:06:17.394074724 +0000 UTC m=+1137.397268575" lastFinishedPulling="2025-12-05 11:06:18.53416012 +0000 UTC m=+1138.537353971" observedRunningTime="2025-12-05 11:06:22.662459813 +0000 UTC m=+1142.665653674" watchObservedRunningTime="2025-12-05 11:06:22.667001386 +0000 UTC m=+1142.670195237" Dec 05 11:06:22.781019 master-0 kubenswrapper[24928]: I1205 11:06:22.780868 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-ec941-backup-0" podStartSLOduration=5.627270896 podStartE2EDuration="6.780851567s" podCreationTimestamp="2025-12-05 11:06:16 +0000 UTC" firstStartedPulling="2025-12-05 11:06:18.389156753 +0000 UTC m=+1138.392350604" lastFinishedPulling="2025-12-05 11:06:19.542737424 +0000 UTC m=+1139.545931275" observedRunningTime="2025-12-05 11:06:22.776027846 +0000 UTC m=+1142.779221717" watchObservedRunningTime="2025-12-05 11:06:22.780851567 +0000 UTC m=+1142.784045408" Dec 05 11:06:23.202569 master-0 kubenswrapper[24928]: I1205 11:06:23.202137 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" podStartSLOduration=5.951267436 podStartE2EDuration="7.202115023s" podCreationTimestamp="2025-12-05 11:06:16 +0000 UTC" firstStartedPulling="2025-12-05 11:06:17.912997267 +0000 UTC m=+1137.916191118" lastFinishedPulling="2025-12-05 11:06:19.163844854 +0000 UTC m=+1139.167038705" observedRunningTime="2025-12-05 11:06:23.193263082 +0000 UTC m=+1143.196456933" watchObservedRunningTime="2025-12-05 11:06:23.202115023 +0000 UTC m=+1143.205308874" Dec 05 11:06:23.256748 master-0 kubenswrapper[24928]: I1205 11:06:23.256555 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7bd67d49b4-fmw2f" podStartSLOduration=5.25653706 podStartE2EDuration="5.25653706s" podCreationTimestamp="2025-12-05 11:06:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:23.229340322 +0000 UTC m=+1143.232534173" watchObservedRunningTime="2025-12-05 11:06:23.25653706 +0000 UTC m=+1143.259730901" Dec 05 11:06:23.290241 master-0 kubenswrapper[24928]: I1205 11:06:23.288613 24928 generic.go:334] "Generic (PLEG): container finished" podID="a9675e13-fa09-4eac-873d-5b8068371c3f" containerID="9149da194cd2fcace386cbee97d69b5b55685679f29ab434fe5ce256a9f87df1" exitCode=0 Dec 05 11:06:23.290241 master-0 kubenswrapper[24928]: I1205 11:06:23.288667 24928 generic.go:334] "Generic (PLEG): container finished" podID="a9675e13-fa09-4eac-873d-5b8068371c3f" containerID="aac6390a98833664c6898788e300db78fe0757cf92c0a297830d83be031a821a" exitCode=143 Dec 05 11:06:23.290241 master-0 kubenswrapper[24928]: I1205 11:06:23.288734 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-api-0" event={"ID":"a9675e13-fa09-4eac-873d-5b8068371c3f","Type":"ContainerDied","Data":"9149da194cd2fcace386cbee97d69b5b55685679f29ab434fe5ce256a9f87df1"} Dec 05 11:06:23.290241 master-0 kubenswrapper[24928]: I1205 11:06:23.288772 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-api-0" event={"ID":"a9675e13-fa09-4eac-873d-5b8068371c3f","Type":"ContainerDied","Data":"aac6390a98833664c6898788e300db78fe0757cf92c0a297830d83be031a821a"} Dec 05 11:06:23.310335 master-0 kubenswrapper[24928]: I1205 11:06:23.306582 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" event={"ID":"c684b0cd-620b-45f3-9fee-4eb6f8cf99df","Type":"ContainerStarted","Data":"37c7fa0465c346fadab41fa46614e8152d9d6def2ed2ac30169efb28ed50a342"} Dec 05 11:06:23.314514 master-0 kubenswrapper[24928]: I1205 11:06:23.310827 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:23.340101 master-0 kubenswrapper[24928]: I1205 11:06:23.337607 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" podStartSLOduration=5.3375878310000004 podStartE2EDuration="5.337587831s" podCreationTimestamp="2025-12-05 11:06:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:23.336781622 +0000 UTC m=+1143.339975473" watchObservedRunningTime="2025-12-05 11:06:23.337587831 +0000 UTC m=+1143.340781682" Dec 05 11:06:23.719784 master-0 kubenswrapper[24928]: I1205 11:06:23.719705 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7c95d944d9-82sjq"] Dec 05 11:06:23.720236 master-0 kubenswrapper[24928]: E1205 11:06:23.720204 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" containerName="init" Dec 05 11:06:23.720236 master-0 kubenswrapper[24928]: I1205 11:06:23.720228 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" containerName="init" Dec 05 11:06:23.720590 master-0 kubenswrapper[24928]: I1205 11:06:23.720560 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" containerName="init" Dec 05 11:06:23.721821 master-0 kubenswrapper[24928]: I1205 11:06:23.721788 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.725833 master-0 kubenswrapper[24928]: I1205 11:06:23.725545 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 05 11:06:23.727850 master-0 kubenswrapper[24928]: I1205 11:06:23.727801 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 05 11:06:23.743237 master-0 kubenswrapper[24928]: I1205 11:06:23.743142 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c95d944d9-82sjq"] Dec 05 11:06:23.837912 master-0 kubenswrapper[24928]: I1205 11:06:23.837841 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79vl5\" (UniqueName: \"kubernetes.io/projected/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-kube-api-access-79vl5\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.838139 master-0 kubenswrapper[24928]: I1205 11:06:23.838014 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-config\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.838139 master-0 kubenswrapper[24928]: I1205 11:06:23.838053 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-internal-tls-certs\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.838139 master-0 kubenswrapper[24928]: I1205 11:06:23.838083 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-ovndb-tls-certs\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.838412 master-0 kubenswrapper[24928]: I1205 11:06:23.838163 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-public-tls-certs\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.838412 master-0 kubenswrapper[24928]: I1205 11:06:23.838190 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-httpd-config\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.838412 master-0 kubenswrapper[24928]: I1205 11:06:23.838209 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-combined-ca-bundle\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.891462 master-0 kubenswrapper[24928]: I1205 11:06:23.889852 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-api-0" Dec 05 11:06:23.947509 master-0 kubenswrapper[24928]: I1205 11:06:23.946943 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data-custom\") pod \"a9675e13-fa09-4eac-873d-5b8068371c3f\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " Dec 05 11:06:23.947509 master-0 kubenswrapper[24928]: I1205 11:06:23.947059 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-scripts\") pod \"a9675e13-fa09-4eac-873d-5b8068371c3f\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " Dec 05 11:06:23.947509 master-0 kubenswrapper[24928]: I1205 11:06:23.947090 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-combined-ca-bundle\") pod \"a9675e13-fa09-4eac-873d-5b8068371c3f\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " Dec 05 11:06:23.947509 master-0 kubenswrapper[24928]: I1205 11:06:23.947150 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a9675e13-fa09-4eac-873d-5b8068371c3f-etc-machine-id\") pod \"a9675e13-fa09-4eac-873d-5b8068371c3f\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " Dec 05 11:06:23.947509 master-0 kubenswrapper[24928]: I1205 11:06:23.947168 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28ssg\" (UniqueName: \"kubernetes.io/projected/a9675e13-fa09-4eac-873d-5b8068371c3f-kube-api-access-28ssg\") pod \"a9675e13-fa09-4eac-873d-5b8068371c3f\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " Dec 05 11:06:23.947509 master-0 kubenswrapper[24928]: I1205 11:06:23.947239 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9675e13-fa09-4eac-873d-5b8068371c3f-logs\") pod \"a9675e13-fa09-4eac-873d-5b8068371c3f\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " Dec 05 11:06:23.947509 master-0 kubenswrapper[24928]: I1205 11:06:23.947402 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data\") pod \"a9675e13-fa09-4eac-873d-5b8068371c3f\" (UID: \"a9675e13-fa09-4eac-873d-5b8068371c3f\") " Dec 05 11:06:23.948045 master-0 kubenswrapper[24928]: I1205 11:06:23.947750 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-public-tls-certs\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.948045 master-0 kubenswrapper[24928]: I1205 11:06:23.947783 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-httpd-config\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.948045 master-0 kubenswrapper[24928]: I1205 11:06:23.947807 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-combined-ca-bundle\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.948045 master-0 kubenswrapper[24928]: I1205 11:06:23.947860 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79vl5\" (UniqueName: \"kubernetes.io/projected/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-kube-api-access-79vl5\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.948045 master-0 kubenswrapper[24928]: I1205 11:06:23.947947 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-config\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.948045 master-0 kubenswrapper[24928]: I1205 11:06:23.947971 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-internal-tls-certs\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.948045 master-0 kubenswrapper[24928]: I1205 11:06:23.947997 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-ovndb-tls-certs\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.952669 master-0 kubenswrapper[24928]: I1205 11:06:23.952510 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9675e13-fa09-4eac-873d-5b8068371c3f-logs" (OuterVolumeSpecName: "logs") pod "a9675e13-fa09-4eac-873d-5b8068371c3f" (UID: "a9675e13-fa09-4eac-873d-5b8068371c3f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:06:23.952986 master-0 kubenswrapper[24928]: I1205 11:06:23.952946 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9675e13-fa09-4eac-873d-5b8068371c3f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a9675e13-fa09-4eac-873d-5b8068371c3f" (UID: "a9675e13-fa09-4eac-873d-5b8068371c3f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:23.955405 master-0 kubenswrapper[24928]: I1205 11:06:23.955335 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a9675e13-fa09-4eac-873d-5b8068371c3f" (UID: "a9675e13-fa09-4eac-873d-5b8068371c3f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:23.955405 master-0 kubenswrapper[24928]: I1205 11:06:23.955397 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-ovndb-tls-certs\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.956578 master-0 kubenswrapper[24928]: I1205 11:06:23.956345 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-httpd-config\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.957877 master-0 kubenswrapper[24928]: I1205 11:06:23.957272 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-internal-tls-certs\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.967729 master-0 kubenswrapper[24928]: I1205 11:06:23.967589 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-config\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.970735 master-0 kubenswrapper[24928]: I1205 11:06:23.970620 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-scripts" (OuterVolumeSpecName: "scripts") pod "a9675e13-fa09-4eac-873d-5b8068371c3f" (UID: "a9675e13-fa09-4eac-873d-5b8068371c3f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:23.971553 master-0 kubenswrapper[24928]: I1205 11:06:23.971508 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-public-tls-certs\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.971788 master-0 kubenswrapper[24928]: I1205 11:06:23.971747 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9675e13-fa09-4eac-873d-5b8068371c3f-kube-api-access-28ssg" (OuterVolumeSpecName: "kube-api-access-28ssg") pod "a9675e13-fa09-4eac-873d-5b8068371c3f" (UID: "a9675e13-fa09-4eac-873d-5b8068371c3f"). InnerVolumeSpecName "kube-api-access-28ssg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:23.979180 master-0 kubenswrapper[24928]: I1205 11:06:23.979129 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79vl5\" (UniqueName: \"kubernetes.io/projected/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-kube-api-access-79vl5\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.982528 master-0 kubenswrapper[24928]: I1205 11:06:23.982480 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40e68c33-2ca2-4b8f-b8e0-0e2dba275a72-combined-ca-bundle\") pod \"neutron-7c95d944d9-82sjq\" (UID: \"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72\") " pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:23.994845 master-0 kubenswrapper[24928]: I1205 11:06:23.994761 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9675e13-fa09-4eac-873d-5b8068371c3f" (UID: "a9675e13-fa09-4eac-873d-5b8068371c3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:24.035178 master-0 kubenswrapper[24928]: I1205 11:06:24.035100 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data" (OuterVolumeSpecName: "config-data") pod "a9675e13-fa09-4eac-873d-5b8068371c3f" (UID: "a9675e13-fa09-4eac-873d-5b8068371c3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:24.058166 master-0 kubenswrapper[24928]: I1205 11:06:24.052123 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:24.058166 master-0 kubenswrapper[24928]: I1205 11:06:24.052196 24928 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-config-data-custom\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:24.058166 master-0 kubenswrapper[24928]: I1205 11:06:24.052209 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:24.058166 master-0 kubenswrapper[24928]: I1205 11:06:24.052220 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9675e13-fa09-4eac-873d-5b8068371c3f-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:24.058166 master-0 kubenswrapper[24928]: I1205 11:06:24.052235 24928 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a9675e13-fa09-4eac-873d-5b8068371c3f-etc-machine-id\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:24.058166 master-0 kubenswrapper[24928]: I1205 11:06:24.052248 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28ssg\" (UniqueName: \"kubernetes.io/projected/a9675e13-fa09-4eac-873d-5b8068371c3f-kube-api-access-28ssg\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:24.058166 master-0 kubenswrapper[24928]: I1205 11:06:24.052261 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9675e13-fa09-4eac-873d-5b8068371c3f-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:24.178060 master-0 kubenswrapper[24928]: I1205 11:06:24.177937 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:24.353270 master-0 kubenswrapper[24928]: I1205 11:06:24.353197 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-api-0" event={"ID":"a9675e13-fa09-4eac-873d-5b8068371c3f","Type":"ContainerDied","Data":"be5a6630c19ef5dd4a1814f19f09a0c93e861d0d9a7a6663e450d230d9055421"} Dec 05 11:06:24.354278 master-0 kubenswrapper[24928]: I1205 11:06:24.353278 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-api-0" Dec 05 11:06:24.354278 master-0 kubenswrapper[24928]: I1205 11:06:24.353285 24928 scope.go:117] "RemoveContainer" containerID="9149da194cd2fcace386cbee97d69b5b55685679f29ab434fe5ce256a9f87df1" Dec 05 11:06:24.392297 master-0 kubenswrapper[24928]: I1205 11:06:24.388521 24928 scope.go:117] "RemoveContainer" containerID="aac6390a98833664c6898788e300db78fe0757cf92c0a297830d83be031a821a" Dec 05 11:06:26.695332 master-0 kubenswrapper[24928]: I1205 11:06:26.695259 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:26.910494 master-0 kubenswrapper[24928]: I1205 11:06:26.910353 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:26.989886 master-0 kubenswrapper[24928]: I1205 11:06:26.989820 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:27.022453 master-0 kubenswrapper[24928]: I1205 11:06:27.021053 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:27.195137 master-0 kubenswrapper[24928]: I1205 11:06:27.195072 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:27.238588 master-0 kubenswrapper[24928]: I1205 11:06:27.238521 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:27.257035 master-0 kubenswrapper[24928]: I1205 11:06:27.256933 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ec941-api-0"] Dec 05 11:06:27.876909 master-0 kubenswrapper[24928]: I1205 11:06:27.876827 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:27.877657 master-0 kubenswrapper[24928]: I1205 11:06:27.877622 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:27.908506 master-0 kubenswrapper[24928]: I1205 11:06:27.908435 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:27.924194 master-0 kubenswrapper[24928]: I1205 11:06:27.924096 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:27.968488 master-0 kubenswrapper[24928]: I1205 11:06:27.968410 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-ec941-api-0"] Dec 05 11:06:28.055713 master-0 kubenswrapper[24928]: I1205 11:06:28.052440 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ec941-api-0"] Dec 05 11:06:28.055713 master-0 kubenswrapper[24928]: E1205 11:06:28.053106 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9675e13-fa09-4eac-873d-5b8068371c3f" containerName="cinder-api" Dec 05 11:06:28.055713 master-0 kubenswrapper[24928]: I1205 11:06:28.053129 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9675e13-fa09-4eac-873d-5b8068371c3f" containerName="cinder-api" Dec 05 11:06:28.055713 master-0 kubenswrapper[24928]: E1205 11:06:28.053148 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9675e13-fa09-4eac-873d-5b8068371c3f" containerName="cinder-ec941-api-log" Dec 05 11:06:28.055713 master-0 kubenswrapper[24928]: I1205 11:06:28.053159 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9675e13-fa09-4eac-873d-5b8068371c3f" containerName="cinder-ec941-api-log" Dec 05 11:06:28.055713 master-0 kubenswrapper[24928]: I1205 11:06:28.053488 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9675e13-fa09-4eac-873d-5b8068371c3f" containerName="cinder-api" Dec 05 11:06:28.055713 master-0 kubenswrapper[24928]: I1205 11:06:28.053525 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9675e13-fa09-4eac-873d-5b8068371c3f" containerName="cinder-ec941-api-log" Dec 05 11:06:28.061559 master-0 kubenswrapper[24928]: I1205 11:06:28.058708 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.063454 master-0 kubenswrapper[24928]: I1205 11:06:28.063362 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ec941-scheduler-0"] Dec 05 11:06:28.064058 master-0 kubenswrapper[24928]: I1205 11:06:28.064004 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 05 11:06:28.064131 master-0 kubenswrapper[24928]: I1205 11:06:28.064092 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-api-config-data" Dec 05 11:06:28.068930 master-0 kubenswrapper[24928]: I1205 11:06:28.068859 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 05 11:06:28.095841 master-0 kubenswrapper[24928]: I1205 11:06:28.095772 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-api-0"] Dec 05 11:06:28.170745 master-0 kubenswrapper[24928]: I1205 11:06:28.170570 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-config-data-custom\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.170745 master-0 kubenswrapper[24928]: I1205 11:06:28.170682 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-scripts\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.171031 master-0 kubenswrapper[24928]: I1205 11:06:28.170940 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4efe64d-2fec-47b4-896a-d144062dcd27-logs\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.171031 master-0 kubenswrapper[24928]: I1205 11:06:28.171006 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-combined-ca-bundle\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.171131 master-0 kubenswrapper[24928]: I1205 11:06:28.171036 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-internal-tls-certs\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.171131 master-0 kubenswrapper[24928]: I1205 11:06:28.171059 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lwbp\" (UniqueName: \"kubernetes.io/projected/a4efe64d-2fec-47b4-896a-d144062dcd27-kube-api-access-9lwbp\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.171131 master-0 kubenswrapper[24928]: I1205 11:06:28.171096 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a4efe64d-2fec-47b4-896a-d144062dcd27-etc-machine-id\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.171275 master-0 kubenswrapper[24928]: I1205 11:06:28.171135 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-public-tls-certs\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.171275 master-0 kubenswrapper[24928]: I1205 11:06:28.171170 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-config-data\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.174453 master-0 kubenswrapper[24928]: I1205 11:06:28.174349 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ec941-backup-0"] Dec 05 11:06:28.235493 master-0 kubenswrapper[24928]: I1205 11:06:28.233746 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9675e13-fa09-4eac-873d-5b8068371c3f" path="/var/lib/kubelet/pods/a9675e13-fa09-4eac-873d-5b8068371c3f/volumes" Dec 05 11:06:28.235493 master-0 kubenswrapper[24928]: I1205 11:06:28.234543 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ec941-volume-lvm-iscsi-0"] Dec 05 11:06:28.241680 master-0 kubenswrapper[24928]: I1205 11:06:28.241614 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7c95d944d9-82sjq"] Dec 05 11:06:28.274632 master-0 kubenswrapper[24928]: I1205 11:06:28.274057 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4efe64d-2fec-47b4-896a-d144062dcd27-logs\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.274632 master-0 kubenswrapper[24928]: I1205 11:06:28.274148 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-combined-ca-bundle\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.274632 master-0 kubenswrapper[24928]: I1205 11:06:28.274174 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-internal-tls-certs\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.274632 master-0 kubenswrapper[24928]: I1205 11:06:28.274199 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lwbp\" (UniqueName: \"kubernetes.io/projected/a4efe64d-2fec-47b4-896a-d144062dcd27-kube-api-access-9lwbp\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.274632 master-0 kubenswrapper[24928]: I1205 11:06:28.274264 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a4efe64d-2fec-47b4-896a-d144062dcd27-etc-machine-id\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.274632 master-0 kubenswrapper[24928]: I1205 11:06:28.274317 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-public-tls-certs\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.274632 master-0 kubenswrapper[24928]: I1205 11:06:28.274344 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-config-data\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.274632 master-0 kubenswrapper[24928]: I1205 11:06:28.274497 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-config-data-custom\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.274632 master-0 kubenswrapper[24928]: I1205 11:06:28.274556 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-scripts\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.275300 master-0 kubenswrapper[24928]: I1205 11:06:28.274939 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a4efe64d-2fec-47b4-896a-d144062dcd27-etc-machine-id\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.283455 master-0 kubenswrapper[24928]: I1205 11:06:28.275873 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4efe64d-2fec-47b4-896a-d144062dcd27-logs\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.283455 master-0 kubenswrapper[24928]: I1205 11:06:28.278321 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-internal-tls-certs\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.283455 master-0 kubenswrapper[24928]: I1205 11:06:28.279310 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-combined-ca-bundle\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.283455 master-0 kubenswrapper[24928]: I1205 11:06:28.279803 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-scripts\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.283455 master-0 kubenswrapper[24928]: I1205 11:06:28.281211 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-public-tls-certs\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.283455 master-0 kubenswrapper[24928]: I1205 11:06:28.281587 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-config-data-custom\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.284457 master-0 kubenswrapper[24928]: I1205 11:06:28.284379 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4efe64d-2fec-47b4-896a-d144062dcd27-config-data\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.301327 master-0 kubenswrapper[24928]: I1205 11:06:28.301258 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lwbp\" (UniqueName: \"kubernetes.io/projected/a4efe64d-2fec-47b4-896a-d144062dcd27-kube-api-access-9lwbp\") pod \"cinder-ec941-api-0\" (UID: \"a4efe64d-2fec-47b4-896a-d144062dcd27\") " pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.402187 master-0 kubenswrapper[24928]: I1205 11:06:28.402117 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-api-0" Dec 05 11:06:28.408543 master-0 kubenswrapper[24928]: I1205 11:06:28.404313 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" podUID="f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" containerName="cinder-volume" containerID="cri-o://489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95" gracePeriod=30 Dec 05 11:06:28.408543 master-0 kubenswrapper[24928]: I1205 11:06:28.404666 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c95d944d9-82sjq" event={"ID":"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72","Type":"ContainerStarted","Data":"fec95198781f79db79f096cf06a3aeb76ff469901356bb857a687b5b96f5a956"} Dec 05 11:06:28.411751 master-0 kubenswrapper[24928]: I1205 11:06:28.411623 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" podUID="f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" containerName="probe" containerID="cri-o://2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a" gracePeriod=30 Dec 05 11:06:28.411874 master-0 kubenswrapper[24928]: I1205 11:06:28.411741 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-ec941-scheduler-0" podUID="6c1e71ab-2601-4cf7-9423-dfb0d1469568" containerName="cinder-scheduler" containerID="cri-o://9ea11651e662d55a392a4b5060336712eeda3be70b16b2e1d9fa94328693f8ab" gracePeriod=30 Dec 05 11:06:28.412057 master-0 kubenswrapper[24928]: I1205 11:06:28.411951 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-ec941-scheduler-0" podUID="6c1e71ab-2601-4cf7-9423-dfb0d1469568" containerName="probe" containerID="cri-o://8acdbec23abbc42cb9d573312a4c8ab4f72724ac083ac0a0d135ed35c957154b" gracePeriod=30 Dec 05 11:06:28.412125 master-0 kubenswrapper[24928]: I1205 11:06:28.412075 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:28.412392 master-0 kubenswrapper[24928]: I1205 11:06:28.412316 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-ec941-backup-0" podUID="9e8aa5cf-e430-4c19-8eca-fe9269299b5b" containerName="cinder-backup" containerID="cri-o://f26dac4d5efacba623abb0a2e976dc18d512c8b3e96f5c357dd52cdef8fa5976" gracePeriod=30 Dec 05 11:06:28.412479 master-0 kubenswrapper[24928]: I1205 11:06:28.412449 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:28.412555 master-0 kubenswrapper[24928]: I1205 11:06:28.412526 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-ec941-backup-0" podUID="9e8aa5cf-e430-4c19-8eca-fe9269299b5b" containerName="probe" containerID="cri-o://b0bfa1ee2a8e48fd72a6d0f74a59625ab9f56c7d9d3527d121ba22124b79d94e" gracePeriod=30 Dec 05 11:06:28.927780 master-0 kubenswrapper[24928]: I1205 11:06:28.927683 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:28.971555 master-0 kubenswrapper[24928]: I1205 11:06:28.971480 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-api-0"] Dec 05 11:06:29.109477 master-0 kubenswrapper[24928]: I1205 11:06:29.076380 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688fbc9bcc-96d2d"] Dec 05 11:06:29.109477 master-0 kubenswrapper[24928]: I1205 11:06:29.076763 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" podUID="628a74ee-1afa-4e41-8b19-c6b0402a3368" containerName="dnsmasq-dns" containerID="cri-o://69a82a43ffa8d33e5be2b5b8b95406f7cecc72aa4a8fec74bdea08272875fac1" gracePeriod=10 Dec 05 11:06:29.460002 master-0 kubenswrapper[24928]: I1205 11:06:29.459904 24928 generic.go:334] "Generic (PLEG): container finished" podID="628a74ee-1afa-4e41-8b19-c6b0402a3368" containerID="69a82a43ffa8d33e5be2b5b8b95406f7cecc72aa4a8fec74bdea08272875fac1" exitCode=0 Dec 05 11:06:29.460190 master-0 kubenswrapper[24928]: I1205 11:06:29.460059 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" event={"ID":"628a74ee-1afa-4e41-8b19-c6b0402a3368","Type":"ContainerDied","Data":"69a82a43ffa8d33e5be2b5b8b95406f7cecc72aa4a8fec74bdea08272875fac1"} Dec 05 11:06:29.461981 master-0 kubenswrapper[24928]: I1205 11:06:29.461921 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-api-0" event={"ID":"a4efe64d-2fec-47b4-896a-d144062dcd27","Type":"ContainerStarted","Data":"5afd4705a3663dcccdce7ac3ac5d2244c451cdcfb91420a64d796fc1a5caea48"} Dec 05 11:06:29.491729 master-0 kubenswrapper[24928]: I1205 11:06:29.491668 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c95d944d9-82sjq" event={"ID":"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72","Type":"ContainerStarted","Data":"1f6ef512ee1f5cf1f047abbd4752acf4391d8295304b650e2d8cf5452124882f"} Dec 05 11:06:29.491729 master-0 kubenswrapper[24928]: I1205 11:06:29.491727 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7c95d944d9-82sjq" event={"ID":"40e68c33-2ca2-4b8f-b8e0-0e2dba275a72","Type":"ContainerStarted","Data":"14b767b92ae88d4f26f983c7b54ba2047cc1e52ea7a07d37e63ebc5988f9b328"} Dec 05 11:06:29.492283 master-0 kubenswrapper[24928]: I1205 11:06:29.492226 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:29.518805 master-0 kubenswrapper[24928]: I1205 11:06:29.515883 24928 generic.go:334] "Generic (PLEG): container finished" podID="9e8aa5cf-e430-4c19-8eca-fe9269299b5b" containerID="b0bfa1ee2a8e48fd72a6d0f74a59625ab9f56c7d9d3527d121ba22124b79d94e" exitCode=0 Dec 05 11:06:29.518805 master-0 kubenswrapper[24928]: I1205 11:06:29.515975 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-backup-0" event={"ID":"9e8aa5cf-e430-4c19-8eca-fe9269299b5b","Type":"ContainerDied","Data":"b0bfa1ee2a8e48fd72a6d0f74a59625ab9f56c7d9d3527d121ba22124b79d94e"} Dec 05 11:06:29.520822 master-0 kubenswrapper[24928]: I1205 11:06:29.520786 24928 generic.go:334] "Generic (PLEG): container finished" podID="f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" containerID="489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95" exitCode=0 Dec 05 11:06:29.521868 master-0 kubenswrapper[24928]: I1205 11:06:29.521834 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" event={"ID":"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b","Type":"ContainerDied","Data":"489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95"} Dec 05 11:06:29.549706 master-0 kubenswrapper[24928]: I1205 11:06:29.549547 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7c95d944d9-82sjq" podStartSLOduration=6.549526155 podStartE2EDuration="6.549526155s" podCreationTimestamp="2025-12-05 11:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:29.54930634 +0000 UTC m=+1149.552500211" watchObservedRunningTime="2025-12-05 11:06:29.549526155 +0000 UTC m=+1149.552720006" Dec 05 11:06:29.980412 master-0 kubenswrapper[24928]: I1205 11:06:29.979943 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:06:30.027689 master-0 kubenswrapper[24928]: I1205 11:06:30.027227 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-sb\") pod \"628a74ee-1afa-4e41-8b19-c6b0402a3368\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " Dec 05 11:06:30.027689 master-0 kubenswrapper[24928]: I1205 11:06:30.027328 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxbth\" (UniqueName: \"kubernetes.io/projected/628a74ee-1afa-4e41-8b19-c6b0402a3368-kube-api-access-vxbth\") pod \"628a74ee-1afa-4e41-8b19-c6b0402a3368\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " Dec 05 11:06:30.027689 master-0 kubenswrapper[24928]: I1205 11:06:30.027445 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-config\") pod \"628a74ee-1afa-4e41-8b19-c6b0402a3368\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " Dec 05 11:06:30.027689 master-0 kubenswrapper[24928]: I1205 11:06:30.027524 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-nb\") pod \"628a74ee-1afa-4e41-8b19-c6b0402a3368\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " Dec 05 11:06:30.027689 master-0 kubenswrapper[24928]: I1205 11:06:30.027679 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-svc\") pod \"628a74ee-1afa-4e41-8b19-c6b0402a3368\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " Dec 05 11:06:30.027689 master-0 kubenswrapper[24928]: I1205 11:06:30.027700 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-swift-storage-0\") pod \"628a74ee-1afa-4e41-8b19-c6b0402a3368\" (UID: \"628a74ee-1afa-4e41-8b19-c6b0402a3368\") " Dec 05 11:06:30.081024 master-0 kubenswrapper[24928]: I1205 11:06:30.080938 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/628a74ee-1afa-4e41-8b19-c6b0402a3368-kube-api-access-vxbth" (OuterVolumeSpecName: "kube-api-access-vxbth") pod "628a74ee-1afa-4e41-8b19-c6b0402a3368" (UID: "628a74ee-1afa-4e41-8b19-c6b0402a3368"). InnerVolumeSpecName "kube-api-access-vxbth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:30.131386 master-0 kubenswrapper[24928]: I1205 11:06:30.131264 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxbth\" (UniqueName: \"kubernetes.io/projected/628a74ee-1afa-4e41-8b19-c6b0402a3368-kube-api-access-vxbth\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.246099 master-0 kubenswrapper[24928]: I1205 11:06:30.246026 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "628a74ee-1afa-4e41-8b19-c6b0402a3368" (UID: "628a74ee-1afa-4e41-8b19-c6b0402a3368"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:30.265533 master-0 kubenswrapper[24928]: I1205 11:06:30.265457 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-config" (OuterVolumeSpecName: "config") pod "628a74ee-1afa-4e41-8b19-c6b0402a3368" (UID: "628a74ee-1afa-4e41-8b19-c6b0402a3368"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:30.269244 master-0 kubenswrapper[24928]: I1205 11:06:30.269182 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "628a74ee-1afa-4e41-8b19-c6b0402a3368" (UID: "628a74ee-1afa-4e41-8b19-c6b0402a3368"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:30.296819 master-0 kubenswrapper[24928]: I1205 11:06:30.296731 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "628a74ee-1afa-4e41-8b19-c6b0402a3368" (UID: "628a74ee-1afa-4e41-8b19-c6b0402a3368"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:30.323306 master-0 kubenswrapper[24928]: I1205 11:06:30.322287 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "628a74ee-1afa-4e41-8b19-c6b0402a3368" (UID: "628a74ee-1afa-4e41-8b19-c6b0402a3368"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:30.337268 master-0 kubenswrapper[24928]: I1205 11:06:30.336188 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.337268 master-0 kubenswrapper[24928]: I1205 11:06:30.337261 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.337636 master-0 kubenswrapper[24928]: I1205 11:06:30.337282 24928 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-dns-swift-storage-0\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.337636 master-0 kubenswrapper[24928]: I1205 11:06:30.337310 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-ovsdbserver-sb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.337636 master-0 kubenswrapper[24928]: I1205 11:06:30.337321 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/628a74ee-1afa-4e41-8b19-c6b0402a3368-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.431976 master-0 kubenswrapper[24928]: I1205 11:06:30.431914 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:30.542556 master-0 kubenswrapper[24928]: I1205 11:06:30.542506 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-combined-ca-bundle\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.542770 master-0 kubenswrapper[24928]: I1205 11:06:30.542580 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-scripts\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.542770 master-0 kubenswrapper[24928]: I1205 11:06:30.542648 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-iscsi\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.542770 master-0 kubenswrapper[24928]: I1205 11:06:30.542698 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.542770 master-0 kubenswrapper[24928]: I1205 11:06:30.542732 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-brick\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.542943 master-0 kubenswrapper[24928]: I1205 11:06:30.542771 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-cinder\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.542943 master-0 kubenswrapper[24928]: I1205 11:06:30.542800 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-lib-modules\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.542943 master-0 kubenswrapper[24928]: I1205 11:06:30.542827 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-machine-id\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.543057 master-0 kubenswrapper[24928]: I1205 11:06:30.542951 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwtpm\" (UniqueName: \"kubernetes.io/projected/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-kube-api-access-bwtpm\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.543057 master-0 kubenswrapper[24928]: I1205 11:06:30.542977 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data-custom\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.543057 master-0 kubenswrapper[24928]: I1205 11:06:30.542996 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-dev\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.543057 master-0 kubenswrapper[24928]: I1205 11:06:30.543051 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-run\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.543212 master-0 kubenswrapper[24928]: I1205 11:06:30.543116 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-lib-cinder\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.543212 master-0 kubenswrapper[24928]: I1205 11:06:30.543152 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-sys\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.543212 master-0 kubenswrapper[24928]: I1205 11:06:30.543168 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-nvme\") pod \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\" (UID: \"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b\") " Dec 05 11:06:30.543698 master-0 kubenswrapper[24928]: I1205 11:06:30.543673 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.543761 master-0 kubenswrapper[24928]: I1205 11:06:30.543708 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.543761 master-0 kubenswrapper[24928]: I1205 11:06:30.543727 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.543761 master-0 kubenswrapper[24928]: I1205 11:06:30.543743 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.543761 master-0 kubenswrapper[24928]: I1205 11:06:30.543760 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.544302 master-0 kubenswrapper[24928]: I1205 11:06:30.544281 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.547056 master-0 kubenswrapper[24928]: I1205 11:06:30.547031 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-scripts" (OuterVolumeSpecName: "scripts") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:30.547387 master-0 kubenswrapper[24928]: I1205 11:06:30.547369 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-run" (OuterVolumeSpecName: "run") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.547579 master-0 kubenswrapper[24928]: I1205 11:06:30.547551 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-kube-api-access-bwtpm" (OuterVolumeSpecName: "kube-api-access-bwtpm") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "kube-api-access-bwtpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:30.547638 master-0 kubenswrapper[24928]: I1205 11:06:30.547595 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.547638 master-0 kubenswrapper[24928]: I1205 11:06:30.547615 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-sys" (OuterVolumeSpecName: "sys") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.547740 master-0 kubenswrapper[24928]: I1205 11:06:30.547637 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-dev" (OuterVolumeSpecName: "dev") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.551290 master-0 kubenswrapper[24928]: I1205 11:06:30.551257 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:30.563281 master-0 kubenswrapper[24928]: I1205 11:06:30.561613 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" event={"ID":"628a74ee-1afa-4e41-8b19-c6b0402a3368","Type":"ContainerDied","Data":"9e8f4e99c9e0dc49a4f7df90a35a56440ae3111a49de2107d58d89aa2a9baad3"} Dec 05 11:06:30.563281 master-0 kubenswrapper[24928]: I1205 11:06:30.561681 24928 scope.go:117] "RemoveContainer" containerID="69a82a43ffa8d33e5be2b5b8b95406f7cecc72aa4a8fec74bdea08272875fac1" Dec 05 11:06:30.563281 master-0 kubenswrapper[24928]: I1205 11:06:30.561834 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688fbc9bcc-96d2d" Dec 05 11:06:30.577774 master-0 kubenswrapper[24928]: I1205 11:06:30.577717 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-backup-0" event={"ID":"9e8aa5cf-e430-4c19-8eca-fe9269299b5b","Type":"ContainerDied","Data":"f26dac4d5efacba623abb0a2e976dc18d512c8b3e96f5c357dd52cdef8fa5976"} Dec 05 11:06:30.577971 master-0 kubenswrapper[24928]: I1205 11:06:30.577651 24928 generic.go:334] "Generic (PLEG): container finished" podID="9e8aa5cf-e430-4c19-8eca-fe9269299b5b" containerID="f26dac4d5efacba623abb0a2e976dc18d512c8b3e96f5c357dd52cdef8fa5976" exitCode=0 Dec 05 11:06:30.580754 master-0 kubenswrapper[24928]: I1205 11:06:30.580720 24928 generic.go:334] "Generic (PLEG): container finished" podID="6c1e71ab-2601-4cf7-9423-dfb0d1469568" containerID="8acdbec23abbc42cb9d573312a4c8ab4f72724ac083ac0a0d135ed35c957154b" exitCode=0 Dec 05 11:06:30.580912 master-0 kubenswrapper[24928]: I1205 11:06:30.580885 24928 generic.go:334] "Generic (PLEG): container finished" podID="6c1e71ab-2601-4cf7-9423-dfb0d1469568" containerID="9ea11651e662d55a392a4b5060336712eeda3be70b16b2e1d9fa94328693f8ab" exitCode=0 Dec 05 11:06:30.581055 master-0 kubenswrapper[24928]: I1205 11:06:30.581024 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-scheduler-0" event={"ID":"6c1e71ab-2601-4cf7-9423-dfb0d1469568","Type":"ContainerDied","Data":"8acdbec23abbc42cb9d573312a4c8ab4f72724ac083ac0a0d135ed35c957154b"} Dec 05 11:06:30.581179 master-0 kubenswrapper[24928]: I1205 11:06:30.581161 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-scheduler-0" event={"ID":"6c1e71ab-2601-4cf7-9423-dfb0d1469568","Type":"ContainerDied","Data":"9ea11651e662d55a392a4b5060336712eeda3be70b16b2e1d9fa94328693f8ab"} Dec 05 11:06:30.583885 master-0 kubenswrapper[24928]: I1205 11:06:30.583861 24928 generic.go:334] "Generic (PLEG): container finished" podID="f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" containerID="2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a" exitCode=0 Dec 05 11:06:30.585495 master-0 kubenswrapper[24928]: I1205 11:06:30.585476 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:30.585981 master-0 kubenswrapper[24928]: I1205 11:06:30.585960 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" event={"ID":"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b","Type":"ContainerDied","Data":"2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a"} Dec 05 11:06:30.586189 master-0 kubenswrapper[24928]: I1205 11:06:30.586168 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" event={"ID":"f499b5c5-b9e1-4609-8dd1-8ce88a97c77b","Type":"ContainerDied","Data":"2a83e06b47b03e0809f8cdbfcd31100606a90720161d5407ad1311a27cd13656"} Dec 05 11:06:30.637922 master-0 kubenswrapper[24928]: I1205 11:06:30.637846 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:30.645822 master-0 kubenswrapper[24928]: I1205 11:06:30.645774 24928 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-brick\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.645822 master-0 kubenswrapper[24928]: I1205 11:06:30.645822 24928 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-locks-cinder\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645834 24928 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-lib-modules\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645849 24928 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-machine-id\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645864 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwtpm\" (UniqueName: \"kubernetes.io/projected/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-kube-api-access-bwtpm\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645877 24928 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data-custom\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645888 24928 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-dev\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645898 24928 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-run\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645909 24928 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-var-lib-cinder\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645919 24928 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-sys\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645930 24928 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-nvme\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645942 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645954 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.646087 master-0 kubenswrapper[24928]: I1205 11:06:30.645965 24928 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-etc-iscsi\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.688894 master-0 kubenswrapper[24928]: I1205 11:06:30.688857 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:30.703700 master-0 kubenswrapper[24928]: I1205 11:06:30.703654 24928 scope.go:117] "RemoveContainer" containerID="1076d61f988b2b7150a031c46ecbb7c89394dc2bfc88106697352ef40c99cbfe" Dec 05 11:06:30.744712 master-0 kubenswrapper[24928]: I1205 11:06:30.744619 24928 scope.go:117] "RemoveContainer" containerID="2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a" Dec 05 11:06:30.745082 master-0 kubenswrapper[24928]: I1205 11:06:30.744691 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data" (OuterVolumeSpecName: "config-data") pod "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" (UID: "f499b5c5-b9e1-4609-8dd1-8ce88a97c77b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:30.749400 master-0 kubenswrapper[24928]: I1205 11:06:30.749358 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.749822 master-0 kubenswrapper[24928]: I1205 11:06:30.749800 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:30.778404 master-0 kubenswrapper[24928]: I1205 11:06:30.778341 24928 scope.go:117] "RemoveContainer" containerID="489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95" Dec 05 11:06:30.805820 master-0 kubenswrapper[24928]: I1205 11:06:30.805752 24928 scope.go:117] "RemoveContainer" containerID="2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a" Dec 05 11:06:30.806892 master-0 kubenswrapper[24928]: E1205 11:06:30.806807 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a\": container with ID starting with 2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a not found: ID does not exist" containerID="2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a" Dec 05 11:06:30.806892 master-0 kubenswrapper[24928]: I1205 11:06:30.806854 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a"} err="failed to get container status \"2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a\": rpc error: code = NotFound desc = could not find container \"2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a\": container with ID starting with 2a8dd40d98fcf2f73677550467eccf341d624e6d8d9a0650624390b47927d39a not found: ID does not exist" Dec 05 11:06:30.806892 master-0 kubenswrapper[24928]: I1205 11:06:30.806893 24928 scope.go:117] "RemoveContainer" containerID="489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95" Dec 05 11:06:30.808085 master-0 kubenswrapper[24928]: E1205 11:06:30.807435 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95\": container with ID starting with 489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95 not found: ID does not exist" containerID="489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95" Dec 05 11:06:30.808085 master-0 kubenswrapper[24928]: I1205 11:06:30.807487 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95"} err="failed to get container status \"489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95\": rpc error: code = NotFound desc = could not find container \"489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95\": container with ID starting with 489f7c8a7c1fe8735b59fe1aec27c995fb8a9867716f5f2902a351ef4d381a95 not found: ID does not exist" Dec 05 11:06:30.851134 master-0 kubenswrapper[24928]: I1205 11:06:30.851072 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5bkl\" (UniqueName: \"kubernetes.io/projected/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-kube-api-access-x5bkl\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.851134 master-0 kubenswrapper[24928]: I1205 11:06:30.851130 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-lib-cinder\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.851437 master-0 kubenswrapper[24928]: I1205 11:06:30.851158 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.851437 master-0 kubenswrapper[24928]: I1205 11:06:30.851241 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-nvme\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.851437 master-0 kubenswrapper[24928]: I1205 11:06:30.851273 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-combined-ca-bundle\") pod \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " Dec 05 11:06:30.851437 master-0 kubenswrapper[24928]: I1205 11:06:30.851313 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-sys\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.851437 master-0 kubenswrapper[24928]: I1205 11:06:30.851375 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-machine-id\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.851437 master-0 kubenswrapper[24928]: I1205 11:06:30.851406 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-iscsi\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.851715 master-0 kubenswrapper[24928]: I1205 11:06:30.851461 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data\") pod \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " Dec 05 11:06:30.851715 master-0 kubenswrapper[24928]: I1205 11:06:30.851489 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c1e71ab-2601-4cf7-9423-dfb0d1469568-etc-machine-id\") pod \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " Dec 05 11:06:30.851715 master-0 kubenswrapper[24928]: I1205 11:06:30.851514 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-brick\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.851715 master-0 kubenswrapper[24928]: I1205 11:06:30.851568 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-run\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.851715 master-0 kubenswrapper[24928]: I1205 11:06:30.851599 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-lib-modules\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.851715 master-0 kubenswrapper[24928]: I1205 11:06:30.851649 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data-custom\") pod \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " Dec 05 11:06:30.851715 master-0 kubenswrapper[24928]: I1205 11:06:30.851687 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-scripts\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.851715 master-0 kubenswrapper[24928]: I1205 11:06:30.851689 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.852050 master-0 kubenswrapper[24928]: I1205 11:06:30.851718 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data-custom\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.852050 master-0 kubenswrapper[24928]: I1205 11:06:30.851800 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.852304 master-0 kubenswrapper[24928]: I1205 11:06:30.852270 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.852373 master-0 kubenswrapper[24928]: I1205 11:06:30.852313 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.852373 master-0 kubenswrapper[24928]: I1205 11:06:30.852336 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.852490 master-0 kubenswrapper[24928]: I1205 11:06:30.852347 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-sys" (OuterVolumeSpecName: "sys") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.852490 master-0 kubenswrapper[24928]: I1205 11:06:30.852444 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c1e71ab-2601-4cf7-9423-dfb0d1469568-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6c1e71ab-2601-4cf7-9423-dfb0d1469568" (UID: "6c1e71ab-2601-4cf7-9423-dfb0d1469568"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.852490 master-0 kubenswrapper[24928]: I1205 11:06:30.852471 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.852653 master-0 kubenswrapper[24928]: I1205 11:06:30.852492 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-run" (OuterVolumeSpecName: "run") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.853746 master-0 kubenswrapper[24928]: I1205 11:06:30.853543 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-scripts\") pod \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " Dec 05 11:06:30.853746 master-0 kubenswrapper[24928]: I1205 11:06:30.853646 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-cinder\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.853746 master-0 kubenswrapper[24928]: I1205 11:06:30.853728 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-combined-ca-bundle\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.853955 master-0 kubenswrapper[24928]: I1205 11:06:30.853757 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lllmr\" (UniqueName: \"kubernetes.io/projected/6c1e71ab-2601-4cf7-9423-dfb0d1469568-kube-api-access-lllmr\") pod \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\" (UID: \"6c1e71ab-2601-4cf7-9423-dfb0d1469568\") " Dec 05 11:06:30.853955 master-0 kubenswrapper[24928]: I1205 11:06:30.853805 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-dev\") pod \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\" (UID: \"9e8aa5cf-e430-4c19-8eca-fe9269299b5b\") " Dec 05 11:06:30.854749 master-0 kubenswrapper[24928]: I1205 11:06:30.854709 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-kube-api-access-x5bkl" (OuterVolumeSpecName: "kube-api-access-x5bkl") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "kube-api-access-x5bkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:30.854828 master-0 kubenswrapper[24928]: I1205 11:06:30.854773 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.854828 master-0 kubenswrapper[24928]: I1205 11:06:30.854798 24928 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6c1e71ab-2601-4cf7-9423-dfb0d1469568-etc-machine-id\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.854828 master-0 kubenswrapper[24928]: I1205 11:06:30.854818 24928 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-brick\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.854828 master-0 kubenswrapper[24928]: I1205 11:06:30.854829 24928 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-run\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.855020 master-0 kubenswrapper[24928]: I1205 11:06:30.854841 24928 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-lib-modules\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.855020 master-0 kubenswrapper[24928]: I1205 11:06:30.854850 24928 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-lib-cinder\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.855020 master-0 kubenswrapper[24928]: I1205 11:06:30.854859 24928 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-nvme\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.855020 master-0 kubenswrapper[24928]: I1205 11:06:30.854867 24928 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-sys\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.855020 master-0 kubenswrapper[24928]: I1205 11:06:30.854875 24928 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-machine-id\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.855020 master-0 kubenswrapper[24928]: I1205 11:06:30.854884 24928 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-etc-iscsi\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.855330 master-0 kubenswrapper[24928]: I1205 11:06:30.855296 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:30.856736 master-0 kubenswrapper[24928]: I1205 11:06:30.856680 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-dev" (OuterVolumeSpecName: "dev") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:06:30.856960 master-0 kubenswrapper[24928]: I1205 11:06:30.856927 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6c1e71ab-2601-4cf7-9423-dfb0d1469568" (UID: "6c1e71ab-2601-4cf7-9423-dfb0d1469568"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:30.858243 master-0 kubenswrapper[24928]: I1205 11:06:30.858203 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-scripts" (OuterVolumeSpecName: "scripts") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:30.862237 master-0 kubenswrapper[24928]: I1205 11:06:30.862201 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-scripts" (OuterVolumeSpecName: "scripts") pod "6c1e71ab-2601-4cf7-9423-dfb0d1469568" (UID: "6c1e71ab-2601-4cf7-9423-dfb0d1469568"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:30.865450 master-0 kubenswrapper[24928]: I1205 11:06:30.865403 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c1e71ab-2601-4cf7-9423-dfb0d1469568-kube-api-access-lllmr" (OuterVolumeSpecName: "kube-api-access-lllmr") pod "6c1e71ab-2601-4cf7-9423-dfb0d1469568" (UID: "6c1e71ab-2601-4cf7-9423-dfb0d1469568"). InnerVolumeSpecName "kube-api-access-lllmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:30.973563 master-0 kubenswrapper[24928]: I1205 11:06:30.973499 24928 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data-custom\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.973563 master-0 kubenswrapper[24928]: I1205 11:06:30.973548 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.973563 master-0 kubenswrapper[24928]: I1205 11:06:30.973559 24928 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data-custom\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.973563 master-0 kubenswrapper[24928]: I1205 11:06:30.973569 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.973563 master-0 kubenswrapper[24928]: I1205 11:06:30.973583 24928 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-var-locks-cinder\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.974058 master-0 kubenswrapper[24928]: I1205 11:06:30.973594 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lllmr\" (UniqueName: \"kubernetes.io/projected/6c1e71ab-2601-4cf7-9423-dfb0d1469568-kube-api-access-lllmr\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.974058 master-0 kubenswrapper[24928]: I1205 11:06:30.973607 24928 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-dev\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.974058 master-0 kubenswrapper[24928]: I1205 11:06:30.973616 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5bkl\" (UniqueName: \"kubernetes.io/projected/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-kube-api-access-x5bkl\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:30.983679 master-0 kubenswrapper[24928]: I1205 11:06:30.983583 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:31.052150 master-0 kubenswrapper[24928]: I1205 11:06:31.052053 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c1e71ab-2601-4cf7-9423-dfb0d1469568" (UID: "6c1e71ab-2601-4cf7-9423-dfb0d1469568"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:31.088317 master-0 kubenswrapper[24928]: I1205 11:06:31.083418 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:31.088317 master-0 kubenswrapper[24928]: I1205 11:06:31.083562 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:06:31.088317 master-0 kubenswrapper[24928]: I1205 11:06:31.083903 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:31.088317 master-0 kubenswrapper[24928]: I1205 11:06:31.083981 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:31.135460 master-0 kubenswrapper[24928]: I1205 11:06:31.135400 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688fbc9bcc-96d2d"] Dec 05 11:06:31.139165 master-0 kubenswrapper[24928]: I1205 11:06:31.138788 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data" (OuterVolumeSpecName: "config-data") pod "6c1e71ab-2601-4cf7-9423-dfb0d1469568" (UID: "6c1e71ab-2601-4cf7-9423-dfb0d1469568"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:31.173451 master-0 kubenswrapper[24928]: I1205 11:06:31.172571 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688fbc9bcc-96d2d"] Dec 05 11:06:31.177975 master-0 kubenswrapper[24928]: I1205 11:06:31.177922 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data" (OuterVolumeSpecName: "config-data") pod "9e8aa5cf-e430-4c19-8eca-fe9269299b5b" (UID: "9e8aa5cf-e430-4c19-8eca-fe9269299b5b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:31.187573 master-0 kubenswrapper[24928]: I1205 11:06:31.187066 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c1e71ab-2601-4cf7-9423-dfb0d1469568-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:31.187573 master-0 kubenswrapper[24928]: I1205 11:06:31.187113 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e8aa5cf-e430-4c19-8eca-fe9269299b5b-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:31.215908 master-0 kubenswrapper[24928]: I1205 11:06:31.198235 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ec941-volume-lvm-iscsi-0"] Dec 05 11:06:31.215908 master-0 kubenswrapper[24928]: I1205 11:06:31.210192 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-ec941-volume-lvm-iscsi-0"] Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.238448 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ec941-volume-lvm-iscsi-0"] Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: E1205 11:06:31.239045 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c1e71ab-2601-4cf7-9423-dfb0d1469568" containerName="probe" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239068 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c1e71ab-2601-4cf7-9423-dfb0d1469568" containerName="probe" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: E1205 11:06:31.239099 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c1e71ab-2601-4cf7-9423-dfb0d1469568" containerName="cinder-scheduler" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239110 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c1e71ab-2601-4cf7-9423-dfb0d1469568" containerName="cinder-scheduler" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: E1205 11:06:31.239130 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8aa5cf-e430-4c19-8eca-fe9269299b5b" containerName="probe" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239139 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8aa5cf-e430-4c19-8eca-fe9269299b5b" containerName="probe" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: E1205 11:06:31.239167 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" containerName="probe" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239177 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" containerName="probe" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: E1205 11:06:31.239209 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="628a74ee-1afa-4e41-8b19-c6b0402a3368" containerName="init" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239218 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="628a74ee-1afa-4e41-8b19-c6b0402a3368" containerName="init" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: E1205 11:06:31.239243 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="628a74ee-1afa-4e41-8b19-c6b0402a3368" containerName="dnsmasq-dns" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239249 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="628a74ee-1afa-4e41-8b19-c6b0402a3368" containerName="dnsmasq-dns" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: E1205 11:06:31.239261 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8aa5cf-e430-4c19-8eca-fe9269299b5b" containerName="cinder-backup" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239267 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8aa5cf-e430-4c19-8eca-fe9269299b5b" containerName="cinder-backup" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: E1205 11:06:31.239291 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" containerName="cinder-volume" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239297 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" containerName="cinder-volume" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239542 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" containerName="probe" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239570 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c1e71ab-2601-4cf7-9423-dfb0d1469568" containerName="cinder-scheduler" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239585 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8aa5cf-e430-4c19-8eca-fe9269299b5b" containerName="cinder-backup" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239608 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="628a74ee-1afa-4e41-8b19-c6b0402a3368" containerName="dnsmasq-dns" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239630 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8aa5cf-e430-4c19-8eca-fe9269299b5b" containerName="probe" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239640 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" containerName="cinder-volume" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.239655 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c1e71ab-2601-4cf7-9423-dfb0d1469568" containerName="probe" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.240997 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.247210 master-0 kubenswrapper[24928]: I1205 11:06:31.244228 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:06:31.253086 master-0 kubenswrapper[24928]: I1205 11:06:31.252804 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-volume-lvm-iscsi-config-data" Dec 05 11:06:31.256742 master-0 kubenswrapper[24928]: I1205 11:06:31.256668 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-volume-lvm-iscsi-0"] Dec 05 11:06:31.392688 master-0 kubenswrapper[24928]: I1205 11:06:31.392639 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-var-locks-brick\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.392894 master-0 kubenswrapper[24928]: I1205 11:06:31.392695 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-lib-modules\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.392894 master-0 kubenswrapper[24928]: I1205 11:06:31.392741 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-config-data\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.392894 master-0 kubenswrapper[24928]: I1205 11:06:31.392768 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-config-data-custom\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.392894 master-0 kubenswrapper[24928]: I1205 11:06:31.392834 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-scripts\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.392894 master-0 kubenswrapper[24928]: I1205 11:06:31.392887 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-dev\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.393137 master-0 kubenswrapper[24928]: I1205 11:06:31.392970 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-var-lib-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.393281 master-0 kubenswrapper[24928]: I1205 11:06:31.393225 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9659\" (UniqueName: \"kubernetes.io/projected/755565d7-e0b1-42cd-a799-27c577b47d70-kube-api-access-h9659\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.393394 master-0 kubenswrapper[24928]: I1205 11:06:31.393363 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-combined-ca-bundle\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.393494 master-0 kubenswrapper[24928]: I1205 11:06:31.393434 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-var-locks-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.393555 master-0 kubenswrapper[24928]: I1205 11:06:31.393495 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-run\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.393609 master-0 kubenswrapper[24928]: I1205 11:06:31.393562 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-etc-machine-id\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.393609 master-0 kubenswrapper[24928]: I1205 11:06:31.393603 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-etc-iscsi\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.393688 master-0 kubenswrapper[24928]: I1205 11:06:31.393627 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-sys\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.393688 master-0 kubenswrapper[24928]: I1205 11:06:31.393665 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-etc-nvme\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.499322 master-0 kubenswrapper[24928]: I1205 11:06:31.499270 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9659\" (UniqueName: \"kubernetes.io/projected/755565d7-e0b1-42cd-a799-27c577b47d70-kube-api-access-h9659\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.499628 master-0 kubenswrapper[24928]: I1205 11:06:31.499601 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-combined-ca-bundle\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.499855 master-0 kubenswrapper[24928]: I1205 11:06:31.499797 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-var-locks-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.500035 master-0 kubenswrapper[24928]: I1205 11:06:31.500014 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-run\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.500209 master-0 kubenswrapper[24928]: I1205 11:06:31.500188 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-etc-machine-id\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.500337 master-0 kubenswrapper[24928]: I1205 11:06:31.500317 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-etc-iscsi\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.500460 master-0 kubenswrapper[24928]: I1205 11:06:31.500441 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-sys\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.500587 master-0 kubenswrapper[24928]: I1205 11:06:31.500568 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-etc-nvme\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.500717 master-0 kubenswrapper[24928]: I1205 11:06:31.500694 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-var-locks-brick\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.500852 master-0 kubenswrapper[24928]: I1205 11:06:31.500803 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-etc-machine-id\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.500914 master-0 kubenswrapper[24928]: I1205 11:06:31.500894 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-var-locks-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.500975 master-0 kubenswrapper[24928]: I1205 11:06:31.500926 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-run\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.501049 master-0 kubenswrapper[24928]: I1205 11:06:31.500830 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-lib-modules\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.501135 master-0 kubenswrapper[24928]: I1205 11:06:31.501110 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-etc-iscsi\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.501274 master-0 kubenswrapper[24928]: I1205 11:06:31.501253 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-config-data\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.501406 master-0 kubenswrapper[24928]: I1205 11:06:31.501386 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-config-data-custom\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.501634 master-0 kubenswrapper[24928]: I1205 11:06:31.501545 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-scripts\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.501775 master-0 kubenswrapper[24928]: I1205 11:06:31.501756 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-dev\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.502030 master-0 kubenswrapper[24928]: I1205 11:06:31.502009 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-var-lib-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.502285 master-0 kubenswrapper[24928]: I1205 11:06:31.502263 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-var-lib-cinder\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.502381 master-0 kubenswrapper[24928]: I1205 11:06:31.501083 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-sys\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.502534 master-0 kubenswrapper[24928]: I1205 11:06:31.502514 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-var-locks-brick\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.502677 master-0 kubenswrapper[24928]: I1205 11:06:31.502657 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-etc-nvme\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.502853 master-0 kubenswrapper[24928]: I1205 11:06:31.502816 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-lib-modules\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.503006 master-0 kubenswrapper[24928]: I1205 11:06:31.502963 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/755565d7-e0b1-42cd-a799-27c577b47d70-dev\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.503550 master-0 kubenswrapper[24928]: I1205 11:06:31.503523 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-combined-ca-bundle\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.506184 master-0 kubenswrapper[24928]: I1205 11:06:31.506150 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-config-data-custom\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.506926 master-0 kubenswrapper[24928]: I1205 11:06:31.506216 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-scripts\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.507251 master-0 kubenswrapper[24928]: I1205 11:06:31.507227 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/755565d7-e0b1-42cd-a799-27c577b47d70-config-data\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.516846 master-0 kubenswrapper[24928]: I1205 11:06:31.516369 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9659\" (UniqueName: \"kubernetes.io/projected/755565d7-e0b1-42cd-a799-27c577b47d70-kube-api-access-h9659\") pod \"cinder-ec941-volume-lvm-iscsi-0\" (UID: \"755565d7-e0b1-42cd-a799-27c577b47d70\") " pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.620576 master-0 kubenswrapper[24928]: I1205 11:06:31.620512 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-api-0" event={"ID":"a4efe64d-2fec-47b4-896a-d144062dcd27","Type":"ContainerStarted","Data":"728b91b652c0e9d3a9994420d538ac67d60e4131a4f795b08d0e12bbecda22f3"} Dec 05 11:06:31.627922 master-0 kubenswrapper[24928]: I1205 11:06:31.625726 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-backup-0" event={"ID":"9e8aa5cf-e430-4c19-8eca-fe9269299b5b","Type":"ContainerDied","Data":"3598120c56fe1399261a2d8197455576335b521ed3a81be8916ef237494a733a"} Dec 05 11:06:31.627922 master-0 kubenswrapper[24928]: I1205 11:06:31.625777 24928 scope.go:117] "RemoveContainer" containerID="b0bfa1ee2a8e48fd72a6d0f74a59625ab9f56c7d9d3527d121ba22124b79d94e" Dec 05 11:06:31.627922 master-0 kubenswrapper[24928]: I1205 11:06:31.626022 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.632891 master-0 kubenswrapper[24928]: I1205 11:06:31.632840 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-scheduler-0" event={"ID":"6c1e71ab-2601-4cf7-9423-dfb0d1469568","Type":"ContainerDied","Data":"137e4aab0a747fe27cee7d03d27e72129c68fc8b5e6cf6ebe66d2947fcbbf24a"} Dec 05 11:06:31.632987 master-0 kubenswrapper[24928]: I1205 11:06:31.632933 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:31.657816 master-0 kubenswrapper[24928]: I1205 11:06:31.657568 24928 scope.go:117] "RemoveContainer" containerID="f26dac4d5efacba623abb0a2e976dc18d512c8b3e96f5c357dd52cdef8fa5976" Dec 05 11:06:31.686892 master-0 kubenswrapper[24928]: I1205 11:06:31.686135 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ec941-backup-0"] Dec 05 11:06:31.717547 master-0 kubenswrapper[24928]: I1205 11:06:31.711077 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:31.729698 master-0 kubenswrapper[24928]: I1205 11:06:31.729553 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-ec941-backup-0"] Dec 05 11:06:31.755633 master-0 kubenswrapper[24928]: I1205 11:06:31.755539 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ec941-scheduler-0"] Dec 05 11:06:31.781227 master-0 kubenswrapper[24928]: I1205 11:06:31.779946 24928 scope.go:117] "RemoveContainer" containerID="8acdbec23abbc42cb9d573312a4c8ab4f72724ac083ac0a0d135ed35c957154b" Dec 05 11:06:31.781227 master-0 kubenswrapper[24928]: I1205 11:06:31.780142 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ec941-backup-0"] Dec 05 11:06:31.782193 master-0 kubenswrapper[24928]: I1205 11:06:31.782155 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.789938 master-0 kubenswrapper[24928]: I1205 11:06:31.789881 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-backup-config-data" Dec 05 11:06:31.804641 master-0 kubenswrapper[24928]: I1205 11:06:31.793643 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-ec941-scheduler-0"] Dec 05 11:06:31.806915 master-0 kubenswrapper[24928]: I1205 11:06:31.806861 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-backup-0"] Dec 05 11:06:31.835364 master-0 kubenswrapper[24928]: I1205 11:06:31.835308 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ec941-scheduler-0"] Dec 05 11:06:31.838509 master-0 kubenswrapper[24928]: I1205 11:06:31.837766 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:31.840654 master-0 kubenswrapper[24928]: I1205 11:06:31.840612 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-ec941-scheduler-config-data" Dec 05 11:06:31.883380 master-0 kubenswrapper[24928]: I1205 11:06:31.883294 24928 scope.go:117] "RemoveContainer" containerID="9ea11651e662d55a392a4b5060336712eeda3be70b16b2e1d9fa94328693f8ab" Dec 05 11:06:31.902503 master-0 kubenswrapper[24928]: I1205 11:06:31.902356 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-scheduler-0"] Dec 05 11:06:31.915842 master-0 kubenswrapper[24928]: I1205 11:06:31.915770 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-var-locks-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916015 master-0 kubenswrapper[24928]: I1205 11:06:31.915848 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-etc-nvme\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916015 master-0 kubenswrapper[24928]: I1205 11:06:31.915907 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-scripts\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916015 master-0 kubenswrapper[24928]: I1205 11:06:31.915934 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mw82\" (UniqueName: \"kubernetes.io/projected/3baefee3-b64c-4eab-bb18-e1ed7516dcda-kube-api-access-8mw82\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916015 master-0 kubenswrapper[24928]: I1205 11:06:31.915963 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-lib-modules\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916015 master-0 kubenswrapper[24928]: I1205 11:06:31.915985 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-etc-iscsi\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916256 master-0 kubenswrapper[24928]: I1205 11:06:31.916031 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-sys\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916256 master-0 kubenswrapper[24928]: I1205 11:06:31.916066 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-var-lib-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916256 master-0 kubenswrapper[24928]: I1205 11:06:31.916086 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-scripts\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:31.916256 master-0 kubenswrapper[24928]: I1205 11:06:31.916115 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-config-data\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916256 master-0 kubenswrapper[24928]: I1205 11:06:31.916160 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtccn\" (UniqueName: \"kubernetes.io/projected/03dad754-1118-40aa-bf45-5abf29614952-kube-api-access-xtccn\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:31.916256 master-0 kubenswrapper[24928]: I1205 11:06:31.916180 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-combined-ca-bundle\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:31.916256 master-0 kubenswrapper[24928]: I1205 11:06:31.916198 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-config-data-custom\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:31.916256 master-0 kubenswrapper[24928]: I1205 11:06:31.916225 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-run\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916256 master-0 kubenswrapper[24928]: I1205 11:06:31.916247 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03dad754-1118-40aa-bf45-5abf29614952-etc-machine-id\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:31.916647 master-0 kubenswrapper[24928]: I1205 11:06:31.916280 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-etc-machine-id\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916647 master-0 kubenswrapper[24928]: I1205 11:06:31.916306 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-config-data-custom\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916647 master-0 kubenswrapper[24928]: I1205 11:06:31.916336 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-var-locks-brick\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916647 master-0 kubenswrapper[24928]: I1205 11:06:31.916353 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-combined-ca-bundle\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:31.916647 master-0 kubenswrapper[24928]: I1205 11:06:31.916379 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-config-data\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:31.916647 master-0 kubenswrapper[24928]: I1205 11:06:31.916402 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-dev\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.019778 master-0 kubenswrapper[24928]: I1205 11:06:32.019651 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-var-locks-brick\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.019778 master-0 kubenswrapper[24928]: I1205 11:06:32.019716 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-combined-ca-bundle\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.019778 master-0 kubenswrapper[24928]: I1205 11:06:32.019761 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-config-data\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.020412 master-0 kubenswrapper[24928]: I1205 11:06:32.019798 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-dev\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.020412 master-0 kubenswrapper[24928]: I1205 11:06:32.019860 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-var-locks-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.020412 master-0 kubenswrapper[24928]: I1205 11:06:32.019900 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-etc-nvme\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.020564 master-0 kubenswrapper[24928]: I1205 11:06:32.020470 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-scripts\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.020564 master-0 kubenswrapper[24928]: I1205 11:06:32.020504 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mw82\" (UniqueName: \"kubernetes.io/projected/3baefee3-b64c-4eab-bb18-e1ed7516dcda-kube-api-access-8mw82\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.020564 master-0 kubenswrapper[24928]: I1205 11:06:32.020546 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-lib-modules\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.020663 master-0 kubenswrapper[24928]: I1205 11:06:32.020566 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-etc-iscsi\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.020663 master-0 kubenswrapper[24928]: I1205 11:06:32.020653 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-sys\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.020723 master-0 kubenswrapper[24928]: I1205 11:06:32.020697 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-var-lib-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.020723 master-0 kubenswrapper[24928]: I1205 11:06:32.020712 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-scripts\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.020793 master-0 kubenswrapper[24928]: I1205 11:06:32.020756 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-config-data\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.021158 master-0 kubenswrapper[24928]: I1205 11:06:32.020833 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtccn\" (UniqueName: \"kubernetes.io/projected/03dad754-1118-40aa-bf45-5abf29614952-kube-api-access-xtccn\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.021158 master-0 kubenswrapper[24928]: I1205 11:06:32.020862 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-combined-ca-bundle\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.021158 master-0 kubenswrapper[24928]: I1205 11:06:32.020882 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-config-data-custom\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.021158 master-0 kubenswrapper[24928]: I1205 11:06:32.020910 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-run\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.021158 master-0 kubenswrapper[24928]: I1205 11:06:32.020930 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03dad754-1118-40aa-bf45-5abf29614952-etc-machine-id\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.021158 master-0 kubenswrapper[24928]: I1205 11:06:32.020967 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-etc-machine-id\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.021158 master-0 kubenswrapper[24928]: I1205 11:06:32.021004 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-config-data-custom\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.024718 master-0 kubenswrapper[24928]: I1205 11:06:32.024652 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-var-locks-brick\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.025180 master-0 kubenswrapper[24928]: I1205 11:06:32.025140 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-config-data-custom\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.025463 master-0 kubenswrapper[24928]: I1205 11:06:32.025402 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-sys\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.029069 master-0 kubenswrapper[24928]: I1205 11:06:32.029019 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-combined-ca-bundle\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.029175 master-0 kubenswrapper[24928]: I1205 11:06:32.029109 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-var-lib-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.031641 master-0 kubenswrapper[24928]: I1205 11:06:32.031589 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-etc-iscsi\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.031855 master-0 kubenswrapper[24928]: I1205 11:06:32.031782 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-config-data\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.032012 master-0 kubenswrapper[24928]: I1205 11:06:32.031949 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-var-locks-cinder\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.032306 master-0 kubenswrapper[24928]: I1205 11:06:32.032242 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-dev\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.033453 master-0 kubenswrapper[24928]: I1205 11:06:32.033348 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-scripts\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.037889 master-0 kubenswrapper[24928]: I1205 11:06:32.037302 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-scripts\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.039671 master-0 kubenswrapper[24928]: I1205 11:06:32.039615 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-lib-modules\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.040065 master-0 kubenswrapper[24928]: I1205 11:06:32.039988 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/03dad754-1118-40aa-bf45-5abf29614952-etc-machine-id\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.040065 master-0 kubenswrapper[24928]: I1205 11:06:32.040054 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-run\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.040378 master-0 kubenswrapper[24928]: I1205 11:06:32.040316 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-etc-machine-id\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.040459 master-0 kubenswrapper[24928]: I1205 11:06:32.040396 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/3baefee3-b64c-4eab-bb18-e1ed7516dcda-etc-nvme\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.044737 master-0 kubenswrapper[24928]: I1205 11:06:32.044363 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-config-data-custom\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.048587 master-0 kubenswrapper[24928]: I1205 11:06:32.047401 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3baefee3-b64c-4eab-bb18-e1ed7516dcda-config-data\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.048587 master-0 kubenswrapper[24928]: I1205 11:06:32.048212 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03dad754-1118-40aa-bf45-5abf29614952-combined-ca-bundle\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.051728 master-0 kubenswrapper[24928]: I1205 11:06:32.051676 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mw82\" (UniqueName: \"kubernetes.io/projected/3baefee3-b64c-4eab-bb18-e1ed7516dcda-kube-api-access-8mw82\") pod \"cinder-ec941-backup-0\" (UID: \"3baefee3-b64c-4eab-bb18-e1ed7516dcda\") " pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.076392 master-0 kubenswrapper[24928]: I1205 11:06:32.060096 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtccn\" (UniqueName: \"kubernetes.io/projected/03dad754-1118-40aa-bf45-5abf29614952-kube-api-access-xtccn\") pod \"cinder-ec941-scheduler-0\" (UID: \"03dad754-1118-40aa-bf45-5abf29614952\") " pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.236262 master-0 kubenswrapper[24928]: I1205 11:06:32.235413 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="628a74ee-1afa-4e41-8b19-c6b0402a3368" path="/var/lib/kubelet/pods/628a74ee-1afa-4e41-8b19-c6b0402a3368/volumes" Dec 05 11:06:32.236841 master-0 kubenswrapper[24928]: I1205 11:06:32.236801 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c1e71ab-2601-4cf7-9423-dfb0d1469568" path="/var/lib/kubelet/pods/6c1e71ab-2601-4cf7-9423-dfb0d1469568/volumes" Dec 05 11:06:32.238285 master-0 kubenswrapper[24928]: I1205 11:06:32.237984 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e8aa5cf-e430-4c19-8eca-fe9269299b5b" path="/var/lib/kubelet/pods/9e8aa5cf-e430-4c19-8eca-fe9269299b5b/volumes" Dec 05 11:06:32.240188 master-0 kubenswrapper[24928]: I1205 11:06:32.240147 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f499b5c5-b9e1-4609-8dd1-8ce88a97c77b" path="/var/lib/kubelet/pods/f499b5c5-b9e1-4609-8dd1-8ce88a97c77b/volumes" Dec 05 11:06:32.298838 master-0 kubenswrapper[24928]: I1205 11:06:32.298144 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:32.333067 master-0 kubenswrapper[24928]: I1205 11:06:32.332994 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:32.409156 master-0 kubenswrapper[24928]: I1205 11:06:32.408677 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-volume-lvm-iscsi-0"] Dec 05 11:06:32.782668 master-0 kubenswrapper[24928]: I1205 11:06:32.782525 24928 generic.go:334] "Generic (PLEG): container finished" podID="0e18fdd2-be67-4228-82a3-4d02d3b350cf" containerID="a33daad2110904d887bdef13f8dcdbc0e41b61ab952bb4b352f57727fbcd6e6e" exitCode=0 Dec 05 11:06:32.782668 master-0 kubenswrapper[24928]: I1205 11:06:32.782617 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-sxns9" event={"ID":"0e18fdd2-be67-4228-82a3-4d02d3b350cf","Type":"ContainerDied","Data":"a33daad2110904d887bdef13f8dcdbc0e41b61ab952bb4b352f57727fbcd6e6e"} Dec 05 11:06:32.788349 master-0 kubenswrapper[24928]: I1205 11:06:32.788293 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-api-0" event={"ID":"a4efe64d-2fec-47b4-896a-d144062dcd27","Type":"ContainerStarted","Data":"c8420350770ec4ddd9e3a66ddea7a12255c17a740c05ee8edf56f71fc6475c0d"} Dec 05 11:06:32.789307 master-0 kubenswrapper[24928]: I1205 11:06:32.789266 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-ec941-api-0" Dec 05 11:06:32.790729 master-0 kubenswrapper[24928]: I1205 11:06:32.790690 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" event={"ID":"755565d7-e0b1-42cd-a799-27c577b47d70","Type":"ContainerStarted","Data":"d2a0db315b7301122faeaae8df29eac569c0cab815e2a182351bccda870d7110"} Dec 05 11:06:32.853195 master-0 kubenswrapper[24928]: I1205 11:06:32.853062 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-ec941-api-0" podStartSLOduration=5.853034208 podStartE2EDuration="5.853034208s" podCreationTimestamp="2025-12-05 11:06:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:32.840590378 +0000 UTC m=+1152.843784229" watchObservedRunningTime="2025-12-05 11:06:32.853034208 +0000 UTC m=+1152.856228059" Dec 05 11:06:33.038968 master-0 kubenswrapper[24928]: I1205 11:06:33.038877 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-backup-0"] Dec 05 11:06:33.056979 master-0 kubenswrapper[24928]: W1205 11:06:33.056527 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3baefee3_b64c_4eab_bb18_e1ed7516dcda.slice/crio-5778dad4815fc649654ee151f8ace662806e94271f97298cd7f1bbcd175fe651 WatchSource:0}: Error finding container 5778dad4815fc649654ee151f8ace662806e94271f97298cd7f1bbcd175fe651: Status 404 returned error can't find the container with id 5778dad4815fc649654ee151f8ace662806e94271f97298cd7f1bbcd175fe651 Dec 05 11:06:33.324529 master-0 kubenswrapper[24928]: I1205 11:06:33.324467 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ec941-scheduler-0"] Dec 05 11:06:33.831684 master-0 kubenswrapper[24928]: I1205 11:06:33.830311 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" event={"ID":"755565d7-e0b1-42cd-a799-27c577b47d70","Type":"ContainerStarted","Data":"0d81de871219d3ad3473fc48e5c29b04e2c715b718e86bee9fc3c754923af47f"} Dec 05 11:06:33.831684 master-0 kubenswrapper[24928]: I1205 11:06:33.830386 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" event={"ID":"755565d7-e0b1-42cd-a799-27c577b47d70","Type":"ContainerStarted","Data":"5833883436d9fde091343351c9b50075f42886d48d951b66de749b011658ef38"} Dec 05 11:06:33.833653 master-0 kubenswrapper[24928]: I1205 11:06:33.833584 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-scheduler-0" event={"ID":"03dad754-1118-40aa-bf45-5abf29614952","Type":"ContainerStarted","Data":"1ede13de9adc43bed36344b0acf2a54d48817efae42f551d85ed138be9ae61aa"} Dec 05 11:06:33.858781 master-0 kubenswrapper[24928]: I1205 11:06:33.858624 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-backup-0" event={"ID":"3baefee3-b64c-4eab-bb18-e1ed7516dcda","Type":"ContainerStarted","Data":"4367c2d7e98c2aa160b2c58d47178948d326a3be5430b545bc86282c3c8a9140"} Dec 05 11:06:33.859131 master-0 kubenswrapper[24928]: I1205 11:06:33.859059 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-backup-0" event={"ID":"3baefee3-b64c-4eab-bb18-e1ed7516dcda","Type":"ContainerStarted","Data":"5778dad4815fc649654ee151f8ace662806e94271f97298cd7f1bbcd175fe651"} Dec 05 11:06:33.861193 master-0 kubenswrapper[24928]: I1205 11:06:33.861050 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" podStartSLOduration=2.861033628 podStartE2EDuration="2.861033628s" podCreationTimestamp="2025-12-05 11:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:33.858077215 +0000 UTC m=+1153.861271066" watchObservedRunningTime="2025-12-05 11:06:33.861033628 +0000 UTC m=+1153.864227479" Dec 05 11:06:34.582871 master-0 kubenswrapper[24928]: I1205 11:06:34.582335 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-sxns9" Dec 05 11:06:34.656126 master-0 kubenswrapper[24928]: I1205 11:06:34.655965 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-combined-ca-bundle\") pod \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " Dec 05 11:06:34.656392 master-0 kubenswrapper[24928]: I1205 11:06:34.656181 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-scripts\") pod \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " Dec 05 11:06:34.656392 master-0 kubenswrapper[24928]: I1205 11:06:34.656276 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data-merged\") pod \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " Dec 05 11:06:34.657285 master-0 kubenswrapper[24928]: I1205 11:06:34.656613 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data\") pod \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " Dec 05 11:06:34.657285 master-0 kubenswrapper[24928]: I1205 11:06:34.656666 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/0e18fdd2-be67-4228-82a3-4d02d3b350cf-etc-podinfo\") pod \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " Dec 05 11:06:34.657285 master-0 kubenswrapper[24928]: I1205 11:06:34.656903 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s48q2\" (UniqueName: \"kubernetes.io/projected/0e18fdd2-be67-4228-82a3-4d02d3b350cf-kube-api-access-s48q2\") pod \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\" (UID: \"0e18fdd2-be67-4228-82a3-4d02d3b350cf\") " Dec 05 11:06:34.663524 master-0 kubenswrapper[24928]: I1205 11:06:34.657008 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "0e18fdd2-be67-4228-82a3-4d02d3b350cf" (UID: "0e18fdd2-be67-4228-82a3-4d02d3b350cf"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:06:34.663524 master-0 kubenswrapper[24928]: I1205 11:06:34.660121 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-scripts" (OuterVolumeSpecName: "scripts") pod "0e18fdd2-be67-4228-82a3-4d02d3b350cf" (UID: "0e18fdd2-be67-4228-82a3-4d02d3b350cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:34.663524 master-0 kubenswrapper[24928]: I1205 11:06:34.663148 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0e18fdd2-be67-4228-82a3-4d02d3b350cf-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "0e18fdd2-be67-4228-82a3-4d02d3b350cf" (UID: "0e18fdd2-be67-4228-82a3-4d02d3b350cf"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 11:06:34.665285 master-0 kubenswrapper[24928]: I1205 11:06:34.664924 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:34.665285 master-0 kubenswrapper[24928]: I1205 11:06:34.665280 24928 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data-merged\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:34.665672 master-0 kubenswrapper[24928]: I1205 11:06:34.665637 24928 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/0e18fdd2-be67-4228-82a3-4d02d3b350cf-etc-podinfo\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:34.702468 master-0 kubenswrapper[24928]: I1205 11:06:34.696065 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e18fdd2-be67-4228-82a3-4d02d3b350cf-kube-api-access-s48q2" (OuterVolumeSpecName: "kube-api-access-s48q2") pod "0e18fdd2-be67-4228-82a3-4d02d3b350cf" (UID: "0e18fdd2-be67-4228-82a3-4d02d3b350cf"). InnerVolumeSpecName "kube-api-access-s48q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:34.761503 master-0 kubenswrapper[24928]: I1205 11:06:34.758578 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data" (OuterVolumeSpecName: "config-data") pod "0e18fdd2-be67-4228-82a3-4d02d3b350cf" (UID: "0e18fdd2-be67-4228-82a3-4d02d3b350cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:34.766307 master-0 kubenswrapper[24928]: I1205 11:06:34.766231 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e18fdd2-be67-4228-82a3-4d02d3b350cf" (UID: "0e18fdd2-be67-4228-82a3-4d02d3b350cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:34.769587 master-0 kubenswrapper[24928]: I1205 11:06:34.768710 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:34.769587 master-0 kubenswrapper[24928]: I1205 11:06:34.768748 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e18fdd2-be67-4228-82a3-4d02d3b350cf-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:34.769587 master-0 kubenswrapper[24928]: I1205 11:06:34.768762 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s48q2\" (UniqueName: \"kubernetes.io/projected/0e18fdd2-be67-4228-82a3-4d02d3b350cf-kube-api-access-s48q2\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:34.928329 master-0 kubenswrapper[24928]: I1205 11:06:34.928124 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-backup-0" event={"ID":"3baefee3-b64c-4eab-bb18-e1ed7516dcda","Type":"ContainerStarted","Data":"089f2b1030e4591a1719df3a6e0077217e4a08010f96998167fb4f8290f4d197"} Dec 05 11:06:34.937511 master-0 kubenswrapper[24928]: I1205 11:06:34.937456 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-sxns9" event={"ID":"0e18fdd2-be67-4228-82a3-4d02d3b350cf","Type":"ContainerDied","Data":"d3a679d6a69ae226505b1d8a46dfc2aff149e33ae9d32741983f6be8bd090375"} Dec 05 11:06:34.937752 master-0 kubenswrapper[24928]: I1205 11:06:34.937736 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3a679d6a69ae226505b1d8a46dfc2aff149e33ae9d32741983f6be8bd090375" Dec 05 11:06:34.937909 master-0 kubenswrapper[24928]: I1205 11:06:34.937895 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-sxns9" Dec 05 11:06:34.949464 master-0 kubenswrapper[24928]: I1205 11:06:34.949373 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-scheduler-0" event={"ID":"03dad754-1118-40aa-bf45-5abf29614952","Type":"ContainerStarted","Data":"66a717236a7321b54141cf1b386b5f98bc148b98b67d86e20bfab8f8a999288e"} Dec 05 11:06:35.039027 master-0 kubenswrapper[24928]: I1205 11:06:35.038892 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-ec941-backup-0" podStartSLOduration=4.038868395 podStartE2EDuration="4.038868395s" podCreationTimestamp="2025-12-05 11:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:35.009993595 +0000 UTC m=+1155.013187446" watchObservedRunningTime="2025-12-05 11:06:35.038868395 +0000 UTC m=+1155.042062246" Dec 05 11:06:35.966177 master-0 kubenswrapper[24928]: I1205 11:06:35.965619 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-neutron-agent-66b77f87b6-l2d4s"] Dec 05 11:06:35.966655 master-0 kubenswrapper[24928]: E1205 11:06:35.966176 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e18fdd2-be67-4228-82a3-4d02d3b350cf" containerName="ironic-db-sync" Dec 05 11:06:35.966655 master-0 kubenswrapper[24928]: I1205 11:06:35.966195 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e18fdd2-be67-4228-82a3-4d02d3b350cf" containerName="ironic-db-sync" Dec 05 11:06:35.966655 master-0 kubenswrapper[24928]: E1205 11:06:35.966241 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e18fdd2-be67-4228-82a3-4d02d3b350cf" containerName="init" Dec 05 11:06:35.966655 master-0 kubenswrapper[24928]: I1205 11:06:35.966247 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e18fdd2-be67-4228-82a3-4d02d3b350cf" containerName="init" Dec 05 11:06:35.966655 master-0 kubenswrapper[24928]: I1205 11:06:35.966491 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e18fdd2-be67-4228-82a3-4d02d3b350cf" containerName="ironic-db-sync" Dec 05 11:06:35.968390 master-0 kubenswrapper[24928]: I1205 11:06:35.968344 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:35.986736 master-0 kubenswrapper[24928]: I1205 11:06:35.980092 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-neutron-agent-config-data" Dec 05 11:06:36.022077 master-0 kubenswrapper[24928]: I1205 11:06:36.019320 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-db-create-z4xfk"] Dec 05 11:06:36.043357 master-0 kubenswrapper[24928]: I1205 11:06:36.042848 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a9f99ee-4977-4963-a6b1-2a2bd2e849d1-combined-ca-bundle\") pod \"ironic-neutron-agent-66b77f87b6-l2d4s\" (UID: \"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1\") " pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:36.043357 master-0 kubenswrapper[24928]: I1205 11:06:36.043077 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82d5g\" (UniqueName: \"kubernetes.io/projected/2a9f99ee-4977-4963-a6b1-2a2bd2e849d1-kube-api-access-82d5g\") pod \"ironic-neutron-agent-66b77f87b6-l2d4s\" (UID: \"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1\") " pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:36.043357 master-0 kubenswrapper[24928]: I1205 11:06:36.043163 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a9f99ee-4977-4963-a6b1-2a2bd2e849d1-config\") pod \"ironic-neutron-agent-66b77f87b6-l2d4s\" (UID: \"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1\") " pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:36.067004 master-0 kubenswrapper[24928]: I1205 11:06:36.048650 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-z4xfk" Dec 05 11:06:36.108482 master-0 kubenswrapper[24928]: I1205 11:06:36.105342 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-neutron-agent-66b77f87b6-l2d4s"] Dec 05 11:06:36.123199 master-0 kubenswrapper[24928]: I1205 11:06:36.123154 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ec941-scheduler-0" event={"ID":"03dad754-1118-40aa-bf45-5abf29614952","Type":"ContainerStarted","Data":"e11573bdf34a85263cdf92186d2b6aeba5984d4829ffc772f54b12058f66d29e"} Dec 05 11:06:36.147160 master-0 kubenswrapper[24928]: I1205 11:06:36.145999 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb7241f8-33c3-4dde-95c8-41a315d34ec8-operator-scripts\") pod \"ironic-inspector-db-create-z4xfk\" (UID: \"eb7241f8-33c3-4dde-95c8-41a315d34ec8\") " pod="openstack/ironic-inspector-db-create-z4xfk" Dec 05 11:06:36.147160 master-0 kubenswrapper[24928]: I1205 11:06:36.146082 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a9f99ee-4977-4963-a6b1-2a2bd2e849d1-config\") pod \"ironic-neutron-agent-66b77f87b6-l2d4s\" (UID: \"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1\") " pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:36.147160 master-0 kubenswrapper[24928]: I1205 11:06:36.146270 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a9f99ee-4977-4963-a6b1-2a2bd2e849d1-combined-ca-bundle\") pod \"ironic-neutron-agent-66b77f87b6-l2d4s\" (UID: \"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1\") " pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:36.147160 master-0 kubenswrapper[24928]: I1205 11:06:36.146716 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plt5m\" (UniqueName: \"kubernetes.io/projected/eb7241f8-33c3-4dde-95c8-41a315d34ec8-kube-api-access-plt5m\") pod \"ironic-inspector-db-create-z4xfk\" (UID: \"eb7241f8-33c3-4dde-95c8-41a315d34ec8\") " pod="openstack/ironic-inspector-db-create-z4xfk" Dec 05 11:06:36.147160 master-0 kubenswrapper[24928]: I1205 11:06:36.146810 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82d5g\" (UniqueName: \"kubernetes.io/projected/2a9f99ee-4977-4963-a6b1-2a2bd2e849d1-kube-api-access-82d5g\") pod \"ironic-neutron-agent-66b77f87b6-l2d4s\" (UID: \"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1\") " pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:36.173516 master-0 kubenswrapper[24928]: I1205 11:06:36.161703 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a9f99ee-4977-4963-a6b1-2a2bd2e849d1-combined-ca-bundle\") pod \"ironic-neutron-agent-66b77f87b6-l2d4s\" (UID: \"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1\") " pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:36.173516 master-0 kubenswrapper[24928]: I1205 11:06:36.166693 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2a9f99ee-4977-4963-a6b1-2a2bd2e849d1-config\") pod \"ironic-neutron-agent-66b77f87b6-l2d4s\" (UID: \"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1\") " pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:36.188921 master-0 kubenswrapper[24928]: I1205 11:06:36.188262 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-create-z4xfk"] Dec 05 11:06:36.194449 master-0 kubenswrapper[24928]: I1205 11:06:36.191703 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82d5g\" (UniqueName: \"kubernetes.io/projected/2a9f99ee-4977-4963-a6b1-2a2bd2e849d1-kube-api-access-82d5g\") pod \"ironic-neutron-agent-66b77f87b6-l2d4s\" (UID: \"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1\") " pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:36.252122 master-0 kubenswrapper[24928]: I1205 11:06:36.249634 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plt5m\" (UniqueName: \"kubernetes.io/projected/eb7241f8-33c3-4dde-95c8-41a315d34ec8-kube-api-access-plt5m\") pod \"ironic-inspector-db-create-z4xfk\" (UID: \"eb7241f8-33c3-4dde-95c8-41a315d34ec8\") " pod="openstack/ironic-inspector-db-create-z4xfk" Dec 05 11:06:36.252122 master-0 kubenswrapper[24928]: I1205 11:06:36.249844 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb7241f8-33c3-4dde-95c8-41a315d34ec8-operator-scripts\") pod \"ironic-inspector-db-create-z4xfk\" (UID: \"eb7241f8-33c3-4dde-95c8-41a315d34ec8\") " pod="openstack/ironic-inspector-db-create-z4xfk" Dec 05 11:06:36.252350 master-0 kubenswrapper[24928]: I1205 11:06:36.252110 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb7241f8-33c3-4dde-95c8-41a315d34ec8-operator-scripts\") pod \"ironic-inspector-db-create-z4xfk\" (UID: \"eb7241f8-33c3-4dde-95c8-41a315d34ec8\") " pod="openstack/ironic-inspector-db-create-z4xfk" Dec 05 11:06:36.256122 master-0 kubenswrapper[24928]: I1205 11:06:36.255886 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-b0db-account-create-update-5x72f"] Dec 05 11:06:36.257327 master-0 kubenswrapper[24928]: I1205 11:06:36.257296 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-b0db-account-create-update-5x72f"] Dec 05 11:06:36.257407 master-0 kubenswrapper[24928]: I1205 11:06:36.257332 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cf777b7bc-z7q5s"] Dec 05 11:06:36.258685 master-0 kubenswrapper[24928]: I1205 11:06:36.258647 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" Dec 05 11:06:36.261447 master-0 kubenswrapper[24928]: I1205 11:06:36.261348 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.261920 master-0 kubenswrapper[24928]: I1205 11:06:36.261871 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-db-secret" Dec 05 11:06:36.277078 master-0 kubenswrapper[24928]: I1205 11:06:36.275695 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cf777b7bc-z7q5s"] Dec 05 11:06:36.300733 master-0 kubenswrapper[24928]: I1205 11:06:36.300696 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plt5m\" (UniqueName: \"kubernetes.io/projected/eb7241f8-33c3-4dde-95c8-41a315d34ec8-kube-api-access-plt5m\") pod \"ironic-inspector-db-create-z4xfk\" (UID: \"eb7241f8-33c3-4dde-95c8-41a315d34ec8\") " pod="openstack/ironic-inspector-db-create-z4xfk" Dec 05 11:06:36.321941 master-0 kubenswrapper[24928]: I1205 11:06:36.320939 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-54d5ff47b5-hzncp"] Dec 05 11:06:36.324454 master-0 kubenswrapper[24928]: I1205 11:06:36.323935 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.333813 master-0 kubenswrapper[24928]: I1205 11:06:36.329234 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-config-data" Dec 05 11:06:36.333813 master-0 kubenswrapper[24928]: I1205 11:06:36.331541 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-transport-url-ironic-transport" Dec 05 11:06:36.333813 master-0 kubenswrapper[24928]: I1205 11:06:36.332182 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-api-scripts" Dec 05 11:06:36.333813 master-0 kubenswrapper[24928]: I1205 11:06:36.332966 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 11:06:36.337864 master-0 kubenswrapper[24928]: I1205 11:06:36.335995 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-api-config-data" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.351253 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-ec941-scheduler-0" podStartSLOduration=5.351228937 podStartE2EDuration="5.351228937s" podCreationTimestamp="2025-12-05 11:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:36.165875483 +0000 UTC m=+1156.169069334" watchObservedRunningTime="2025-12-05 11:06:36.351228937 +0000 UTC m=+1156.354422818" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.353734 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwz2z\" (UniqueName: \"kubernetes.io/projected/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-kube-api-access-fwz2z\") pod \"ironic-inspector-b0db-account-create-update-5x72f\" (UID: \"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa\") " pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.353854 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-custom\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.354209 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-config\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.354310 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-swift-storage-0\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.354332 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.354371 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-merged\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.354496 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pzjc\" (UniqueName: \"kubernetes.io/projected/635e41cc-7a74-43a1-b4c8-2b069496dc3a-kube-api-access-2pzjc\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.354561 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7xts\" (UniqueName: \"kubernetes.io/projected/f4540893-bc29-4b48-a741-ecb016eb0e73-kube-api-access-s7xts\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.354622 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f4540893-bc29-4b48-a741-ecb016eb0e73-etc-podinfo\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.354676 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-scripts\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.354729 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-svc\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.354878 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-operator-scripts\") pod \"ironic-inspector-b0db-account-create-update-5x72f\" (UID: \"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa\") " pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.355025 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.355052 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-combined-ca-bundle\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.355076 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-logs\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.365444 master-0 kubenswrapper[24928]: I1205 11:06:36.355106 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.373454 master-0 kubenswrapper[24928]: I1205 11:06:36.370267 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-54d5ff47b5-hzncp"] Dec 05 11:06:36.390206 master-0 kubenswrapper[24928]: I1205 11:06:36.390157 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:36.457754 master-0 kubenswrapper[24928]: I1205 11:06:36.457623 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f4540893-bc29-4b48-a741-ecb016eb0e73-etc-podinfo\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.457754 master-0 kubenswrapper[24928]: I1205 11:06:36.457731 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-scripts\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.458040 master-0 kubenswrapper[24928]: I1205 11:06:36.457950 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-svc\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.458163 master-0 kubenswrapper[24928]: I1205 11:06:36.458121 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-operator-scripts\") pod \"ironic-inspector-b0db-account-create-update-5x72f\" (UID: \"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa\") " pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" Dec 05 11:06:36.458691 master-0 kubenswrapper[24928]: I1205 11:06:36.458291 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.458850 master-0 kubenswrapper[24928]: I1205 11:06:36.458812 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-combined-ca-bundle\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.458909 master-0 kubenswrapper[24928]: I1205 11:06:36.458856 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-logs\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.459096 master-0 kubenswrapper[24928]: I1205 11:06:36.458917 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.459096 master-0 kubenswrapper[24928]: I1205 11:06:36.459085 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwz2z\" (UniqueName: \"kubernetes.io/projected/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-kube-api-access-fwz2z\") pod \"ironic-inspector-b0db-account-create-update-5x72f\" (UID: \"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa\") " pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" Dec 05 11:06:36.459637 master-0 kubenswrapper[24928]: I1205 11:06:36.459257 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-custom\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.459799 master-0 kubenswrapper[24928]: I1205 11:06:36.459764 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-config\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.459858 master-0 kubenswrapper[24928]: I1205 11:06:36.459811 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-swift-storage-0\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.459903 master-0 kubenswrapper[24928]: I1205 11:06:36.459864 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.459970 master-0 kubenswrapper[24928]: I1205 11:06:36.459944 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-merged\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.460054 master-0 kubenswrapper[24928]: I1205 11:06:36.460031 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pzjc\" (UniqueName: \"kubernetes.io/projected/635e41cc-7a74-43a1-b4c8-2b069496dc3a-kube-api-access-2pzjc\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.460142 master-0 kubenswrapper[24928]: I1205 11:06:36.460115 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7xts\" (UniqueName: \"kubernetes.io/projected/f4540893-bc29-4b48-a741-ecb016eb0e73-kube-api-access-s7xts\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.477459 master-0 kubenswrapper[24928]: I1205 11:06:36.477377 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f4540893-bc29-4b48-a741-ecb016eb0e73-etc-podinfo\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.478583 master-0 kubenswrapper[24928]: I1205 11:06:36.478542 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-nb\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.479529 master-0 kubenswrapper[24928]: I1205 11:06:36.479487 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-operator-scripts\") pod \"ironic-inspector-b0db-account-create-update-5x72f\" (UID: \"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa\") " pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" Dec 05 11:06:36.480307 master-0 kubenswrapper[24928]: I1205 11:06:36.480233 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-swift-storage-0\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.480728 master-0 kubenswrapper[24928]: I1205 11:06:36.480696 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-logs\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.481285 master-0 kubenswrapper[24928]: I1205 11:06:36.481247 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-config\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.482172 master-0 kubenswrapper[24928]: I1205 11:06:36.482129 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-svc\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.482490 master-0 kubenswrapper[24928]: I1205 11:06:36.482405 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-sb\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.483820 master-0 kubenswrapper[24928]: I1205 11:06:36.483507 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-custom\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.485369 master-0 kubenswrapper[24928]: I1205 11:06:36.484842 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-merged\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.489023 master-0 kubenswrapper[24928]: I1205 11:06:36.488963 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.512693 master-0 kubenswrapper[24928]: I1205 11:06:36.497174 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7xts\" (UniqueName: \"kubernetes.io/projected/f4540893-bc29-4b48-a741-ecb016eb0e73-kube-api-access-s7xts\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.512693 master-0 kubenswrapper[24928]: I1205 11:06:36.502888 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-combined-ca-bundle\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.521301 master-0 kubenswrapper[24928]: I1205 11:06:36.521242 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pzjc\" (UniqueName: \"kubernetes.io/projected/635e41cc-7a74-43a1-b4c8-2b069496dc3a-kube-api-access-2pzjc\") pod \"dnsmasq-dns-7cf777b7bc-z7q5s\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.522793 master-0 kubenswrapper[24928]: I1205 11:06:36.522705 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-scripts\") pod \"ironic-54d5ff47b5-hzncp\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.526275 master-0 kubenswrapper[24928]: I1205 11:06:36.526207 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwz2z\" (UniqueName: \"kubernetes.io/projected/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-kube-api-access-fwz2z\") pod \"ironic-inspector-b0db-account-create-update-5x72f\" (UID: \"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa\") " pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" Dec 05 11:06:36.558527 master-0 kubenswrapper[24928]: I1205 11:06:36.552999 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-z4xfk" Dec 05 11:06:36.635293 master-0 kubenswrapper[24928]: I1205 11:06:36.635195 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" Dec 05 11:06:36.672937 master-0 kubenswrapper[24928]: I1205 11:06:36.671392 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:36.687703 master-0 kubenswrapper[24928]: I1205 11:06:36.687444 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:36.711758 master-0 kubenswrapper[24928]: I1205 11:06:36.711688 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:36.835263 master-0 kubenswrapper[24928]: I1205 11:06:36.835198 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6dc4d774b6-b9nzv" Dec 05 11:06:37.028234 master-0 kubenswrapper[24928]: I1205 11:06:37.016359 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-neutron-agent-66b77f87b6-l2d4s"] Dec 05 11:06:37.193055 master-0 kubenswrapper[24928]: I1205 11:06:37.192976 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" event={"ID":"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1","Type":"ContainerStarted","Data":"43b4a9ce1f16d5fc8987d224580c187d2f0b7d2fc994502ade82e6de81c8f03f"} Dec 05 11:06:37.317175 master-0 kubenswrapper[24928]: I1205 11:06:37.316772 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:37.339936 master-0 kubenswrapper[24928]: I1205 11:06:37.339882 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:37.625765 master-0 kubenswrapper[24928]: I1205 11:06:37.610023 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-create-z4xfk"] Dec 05 11:06:37.656212 master-0 kubenswrapper[24928]: I1205 11:06:37.656106 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-b0db-account-create-update-5x72f"] Dec 05 11:06:37.782905 master-0 kubenswrapper[24928]: I1205 11:06:37.782029 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cf777b7bc-z7q5s"] Dec 05 11:06:37.784968 master-0 kubenswrapper[24928]: W1205 11:06:37.784796 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod635e41cc_7a74_43a1_b4c8_2b069496dc3a.slice/crio-094ae3dc207da40dce773f9a423aef73ae6d39279ba3830995ec7f15c8805c38 WatchSource:0}: Error finding container 094ae3dc207da40dce773f9a423aef73ae6d39279ba3830995ec7f15c8805c38: Status 404 returned error can't find the container with id 094ae3dc207da40dce773f9a423aef73ae6d39279ba3830995ec7f15c8805c38 Dec 05 11:06:38.187885 master-0 kubenswrapper[24928]: I1205 11:06:38.187810 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-54d5ff47b5-hzncp"] Dec 05 11:06:38.211458 master-0 kubenswrapper[24928]: I1205 11:06:38.201179 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-conductor-0"] Dec 05 11:06:38.240040 master-0 kubenswrapper[24928]: I1205 11:06:38.234006 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-conductor-0" Dec 05 11:06:38.240273 master-0 kubenswrapper[24928]: I1205 11:06:38.240058 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-conductor-scripts" Dec 05 11:06:38.240323 master-0 kubenswrapper[24928]: I1205 11:06:38.240312 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-conductor-config-data" Dec 05 11:06:38.365385 master-0 kubenswrapper[24928]: I1205 11:06:38.362470 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-conductor-0"] Dec 05 11:06:38.365385 master-0 kubenswrapper[24928]: I1205 11:06:38.364551 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-z4xfk" event={"ID":"eb7241f8-33c3-4dde-95c8-41a315d34ec8","Type":"ContainerStarted","Data":"f599b091f4901fcc1add1364c893008e0c0791f7fd67d295d56483df969df21d"} Dec 05 11:06:38.365385 master-0 kubenswrapper[24928]: I1205 11:06:38.364580 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-z4xfk" event={"ID":"eb7241f8-33c3-4dde-95c8-41a315d34ec8","Type":"ContainerStarted","Data":"e8a0dbd47040de38459413107627b2dac54b98ec077562a47a530a87c06abb19"} Dec 05 11:06:38.372567 master-0 kubenswrapper[24928]: I1205 11:06:38.372489 24928 generic.go:334] "Generic (PLEG): container finished" podID="635e41cc-7a74-43a1-b4c8-2b069496dc3a" containerID="cca05638df01834d73cf5f18ee70814cf850b91358bd23f1a178f51e9764d727" exitCode=0 Dec 05 11:06:38.372969 master-0 kubenswrapper[24928]: I1205 11:06:38.372942 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" event={"ID":"635e41cc-7a74-43a1-b4c8-2b069496dc3a","Type":"ContainerDied","Data":"cca05638df01834d73cf5f18ee70814cf850b91358bd23f1a178f51e9764d727"} Dec 05 11:06:38.373130 master-0 kubenswrapper[24928]: I1205 11:06:38.373110 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" event={"ID":"635e41cc-7a74-43a1-b4c8-2b069496dc3a","Type":"ContainerStarted","Data":"094ae3dc207da40dce773f9a423aef73ae6d39279ba3830995ec7f15c8805c38"} Dec 05 11:06:38.392738 master-0 kubenswrapper[24928]: I1205 11:06:38.389816 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" event={"ID":"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa","Type":"ContainerStarted","Data":"faa12fcdcf2ef663a50455c97db88d084a0ba62848d5c28f9330b220ff68e32d"} Dec 05 11:06:38.392738 master-0 kubenswrapper[24928]: I1205 11:06:38.389862 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" event={"ID":"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa","Type":"ContainerStarted","Data":"6f1cd01f60de8c2e1c64b4fac2a91464250f4f60da94d08407b622a93785f63e"} Dec 05 11:06:38.413147 master-0 kubenswrapper[24928]: I1205 11:06:38.411466 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-db-create-z4xfk" podStartSLOduration=3.411449094 podStartE2EDuration="3.411449094s" podCreationTimestamp="2025-12-05 11:06:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:38.406790281 +0000 UTC m=+1158.409984162" watchObservedRunningTime="2025-12-05 11:06:38.411449094 +0000 UTC m=+1158.414642955" Dec 05 11:06:38.415404 master-0 kubenswrapper[24928]: I1205 11:06:38.414524 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.415404 master-0 kubenswrapper[24928]: I1205 11:06:38.414591 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9djz\" (UniqueName: \"kubernetes.io/projected/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-kube-api-access-d9djz\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.415404 master-0 kubenswrapper[24928]: I1205 11:06:38.414617 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-29e15954-fce3-40d2-8a78-38683d7219c1\" (UniqueName: \"kubernetes.io/csi/topolvm.io^f413b605-6ed5-4ee6-aee2-a5bb9f64efaf\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.415404 master-0 kubenswrapper[24928]: I1205 11:06:38.414650 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.415404 master-0 kubenswrapper[24928]: I1205 11:06:38.414691 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-config-data\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.415404 master-0 kubenswrapper[24928]: I1205 11:06:38.414724 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.415404 master-0 kubenswrapper[24928]: I1205 11:06:38.414756 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-scripts\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.415404 master-0 kubenswrapper[24928]: I1205 11:06:38.414960 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.506455 master-0 kubenswrapper[24928]: I1205 11:06:38.498466 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" podStartSLOduration=3.498438126 podStartE2EDuration="3.498438126s" podCreationTimestamp="2025-12-05 11:06:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:38.488158725 +0000 UTC m=+1158.491352586" watchObservedRunningTime="2025-12-05 11:06:38.498438126 +0000 UTC m=+1158.501631977" Dec 05 11:06:38.532567 master-0 kubenswrapper[24928]: I1205 11:06:38.529949 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.532567 master-0 kubenswrapper[24928]: I1205 11:06:38.530027 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9djz\" (UniqueName: \"kubernetes.io/projected/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-kube-api-access-d9djz\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.532567 master-0 kubenswrapper[24928]: I1205 11:06:38.530343 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-29e15954-fce3-40d2-8a78-38683d7219c1\" (UniqueName: \"kubernetes.io/csi/topolvm.io^f413b605-6ed5-4ee6-aee2-a5bb9f64efaf\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.532567 master-0 kubenswrapper[24928]: I1205 11:06:38.530901 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.532567 master-0 kubenswrapper[24928]: I1205 11:06:38.531075 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-config-data\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.532567 master-0 kubenswrapper[24928]: I1205 11:06:38.531156 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.532567 master-0 kubenswrapper[24928]: I1205 11:06:38.531242 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-scripts\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.532567 master-0 kubenswrapper[24928]: I1205 11:06:38.531842 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.536621 master-0 kubenswrapper[24928]: I1205 11:06:38.533904 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:06:38.536621 master-0 kubenswrapper[24928]: I1205 11:06:38.533945 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-29e15954-fce3-40d2-8a78-38683d7219c1\" (UniqueName: \"kubernetes.io/csi/topolvm.io^f413b605-6ed5-4ee6-aee2-a5bb9f64efaf\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/910e2211d1bceef52f39d9f0ff89ea93faf86c77fe0c5ced19ffd9e7bd238f7f/globalmount\"" pod="openstack/ironic-conductor-0" Dec 05 11:06:38.536621 master-0 kubenswrapper[24928]: I1205 11:06:38.533950 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.536621 master-0 kubenswrapper[24928]: I1205 11:06:38.534367 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.536621 master-0 kubenswrapper[24928]: I1205 11:06:38.535519 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-scripts\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.541111 master-0 kubenswrapper[24928]: I1205 11:06:38.540520 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.541243 master-0 kubenswrapper[24928]: I1205 11:06:38.541150 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-config-data\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.553773 master-0 kubenswrapper[24928]: I1205 11:06:38.553694 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9djz\" (UniqueName: \"kubernetes.io/projected/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-kube-api-access-d9djz\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:38.562132 master-0 kubenswrapper[24928]: I1205 11:06:38.558657 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d8079b6-0de1-4ffd-b008-9d74bf9673ae-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:39.424618 master-0 kubenswrapper[24928]: I1205 11:06:39.420840 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-54d5ff47b5-hzncp" event={"ID":"f4540893-bc29-4b48-a741-ecb016eb0e73","Type":"ContainerStarted","Data":"40de73483f3d37ce317b8d6af0082084388fd8d7a2b1ae8a16b5008f83dc874f"} Dec 05 11:06:39.425724 master-0 kubenswrapper[24928]: I1205 11:06:39.425645 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" event={"ID":"635e41cc-7a74-43a1-b4c8-2b069496dc3a","Type":"ContainerStarted","Data":"99bfdf0f903af317086040622d17c3c36a8d8bb525ea8a062a787c708b7db1c7"} Dec 05 11:06:39.425812 master-0 kubenswrapper[24928]: I1205 11:06:39.425743 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:39.431503 master-0 kubenswrapper[24928]: I1205 11:06:39.431402 24928 generic.go:334] "Generic (PLEG): container finished" podID="55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa" containerID="faa12fcdcf2ef663a50455c97db88d084a0ba62848d5c28f9330b220ff68e32d" exitCode=0 Dec 05 11:06:39.431699 master-0 kubenswrapper[24928]: I1205 11:06:39.431602 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" event={"ID":"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa","Type":"ContainerDied","Data":"faa12fcdcf2ef663a50455c97db88d084a0ba62848d5c28f9330b220ff68e32d"} Dec 05 11:06:39.434834 master-0 kubenswrapper[24928]: I1205 11:06:39.434309 24928 generic.go:334] "Generic (PLEG): container finished" podID="eb7241f8-33c3-4dde-95c8-41a315d34ec8" containerID="f599b091f4901fcc1add1364c893008e0c0791f7fd67d295d56483df969df21d" exitCode=0 Dec 05 11:06:39.434834 master-0 kubenswrapper[24928]: I1205 11:06:39.434380 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-z4xfk" event={"ID":"eb7241f8-33c3-4dde-95c8-41a315d34ec8","Type":"ContainerDied","Data":"f599b091f4901fcc1add1364c893008e0c0791f7fd67d295d56483df969df21d"} Dec 05 11:06:39.458130 master-0 kubenswrapper[24928]: I1205 11:06:39.458026 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" podStartSLOduration=4.457998069 podStartE2EDuration="4.457998069s" podCreationTimestamp="2025-12-05 11:06:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:39.444769516 +0000 UTC m=+1159.447963367" watchObservedRunningTime="2025-12-05 11:06:39.457998069 +0000 UTC m=+1159.461191930" Dec 05 11:06:39.864532 master-0 kubenswrapper[24928]: I1205 11:06:39.862780 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-647984b79f-nk9b6"] Dec 05 11:06:39.867157 master-0 kubenswrapper[24928]: I1205 11:06:39.866110 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:39.894819 master-0 kubenswrapper[24928]: I1205 11:06:39.893370 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-internal-svc" Dec 05 11:06:39.894819 master-0 kubenswrapper[24928]: I1205 11:06:39.894445 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-public-svc" Dec 05 11:06:39.895929 master-0 kubenswrapper[24928]: I1205 11:06:39.895649 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-647984b79f-nk9b6"] Dec 05 11:06:39.910697 master-0 kubenswrapper[24928]: I1205 11:06:39.898648 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d3f7004-5e71-452f-8abd-907e10bee262-etc-podinfo\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:39.910697 master-0 kubenswrapper[24928]: I1205 11:06:39.898798 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-combined-ca-bundle\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:39.910697 master-0 kubenswrapper[24928]: I1205 11:06:39.898915 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-internal-tls-certs\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:39.910697 master-0 kubenswrapper[24928]: I1205 11:06:39.898956 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1d3f7004-5e71-452f-8abd-907e10bee262-config-data-merged\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:39.910697 master-0 kubenswrapper[24928]: I1205 11:06:39.898992 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-public-tls-certs\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:39.910697 master-0 kubenswrapper[24928]: I1205 11:06:39.899030 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-scripts\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:39.910697 master-0 kubenswrapper[24928]: I1205 11:06:39.899124 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-config-data\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:39.910697 master-0 kubenswrapper[24928]: I1205 11:06:39.899185 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d3f7004-5e71-452f-8abd-907e10bee262-logs\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:39.910697 master-0 kubenswrapper[24928]: I1205 11:06:39.899263 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-config-data-custom\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:39.910697 master-0 kubenswrapper[24928]: I1205 11:06:39.899320 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq45z\" (UniqueName: \"kubernetes.io/projected/1d3f7004-5e71-452f-8abd-907e10bee262-kube-api-access-zq45z\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.004048 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-internal-tls-certs\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.004290 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1d3f7004-5e71-452f-8abd-907e10bee262-config-data-merged\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.004338 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-public-tls-certs\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.004510 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-scripts\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.004653 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-config-data\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.004723 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d3f7004-5e71-452f-8abd-907e10bee262-logs\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.004757 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1d3f7004-5e71-452f-8abd-907e10bee262-config-data-merged\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.004808 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-config-data-custom\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.004873 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq45z\" (UniqueName: \"kubernetes.io/projected/1d3f7004-5e71-452f-8abd-907e10bee262-kube-api-access-zq45z\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.005164 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d3f7004-5e71-452f-8abd-907e10bee262-etc-podinfo\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.005304 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-combined-ca-bundle\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.006767 master-0 kubenswrapper[24928]: I1205 11:06:40.005529 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1d3f7004-5e71-452f-8abd-907e10bee262-logs\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.010756 master-0 kubenswrapper[24928]: I1205 11:06:40.010675 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-combined-ca-bundle\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.010967 master-0 kubenswrapper[24928]: I1205 11:06:40.010815 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-config-data\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.010967 master-0 kubenswrapper[24928]: I1205 11:06:40.010826 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d3f7004-5e71-452f-8abd-907e10bee262-etc-podinfo\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.011495 master-0 kubenswrapper[24928]: I1205 11:06:40.011438 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-scripts\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.013208 master-0 kubenswrapper[24928]: I1205 11:06:40.013154 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-internal-tls-certs\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.019225 master-0 kubenswrapper[24928]: I1205 11:06:40.019167 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-config-data-custom\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.032409 master-0 kubenswrapper[24928]: I1205 11:06:40.030161 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-29e15954-fce3-40d2-8a78-38683d7219c1\" (UniqueName: \"kubernetes.io/csi/topolvm.io^f413b605-6ed5-4ee6-aee2-a5bb9f64efaf\") pod \"ironic-conductor-0\" (UID: \"2d8079b6-0de1-4ffd-b008-9d74bf9673ae\") " pod="openstack/ironic-conductor-0" Dec 05 11:06:40.032409 master-0 kubenswrapper[24928]: I1205 11:06:40.031875 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1d3f7004-5e71-452f-8abd-907e10bee262-public-tls-certs\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.032409 master-0 kubenswrapper[24928]: I1205 11:06:40.032125 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq45z\" (UniqueName: \"kubernetes.io/projected/1d3f7004-5e71-452f-8abd-907e10bee262-kube-api-access-zq45z\") pod \"ironic-647984b79f-nk9b6\" (UID: \"1d3f7004-5e71-452f-8abd-907e10bee262\") " pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.159615 master-0 kubenswrapper[24928]: I1205 11:06:40.159497 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-conductor-0" Dec 05 11:06:40.194467 master-0 kubenswrapper[24928]: I1205 11:06:40.193970 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:40.772990 master-0 kubenswrapper[24928]: I1205 11:06:40.772942 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-ec941-api-0" Dec 05 11:06:41.753540 master-0 kubenswrapper[24928]: I1205 11:06:41.747572 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-647984b79f-nk9b6"] Dec 05 11:06:41.785878 master-0 kubenswrapper[24928]: W1205 11:06:41.785827 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d3f7004_5e71_452f_8abd_907e10bee262.slice/crio-3be063f5d5616568b143ef9f671a0fa79bf310ca3a1d34a11816cd2df6467379 WatchSource:0}: Error finding container 3be063f5d5616568b143ef9f671a0fa79bf310ca3a1d34a11816cd2df6467379: Status 404 returned error can't find the container with id 3be063f5d5616568b143ef9f671a0fa79bf310ca3a1d34a11816cd2df6467379 Dec 05 11:06:41.892274 master-0 kubenswrapper[24928]: I1205 11:06:41.892225 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-z4xfk" Dec 05 11:06:41.905470 master-0 kubenswrapper[24928]: I1205 11:06:41.904978 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" Dec 05 11:06:42.009451 master-0 kubenswrapper[24928]: I1205 11:06:42.009131 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb7241f8-33c3-4dde-95c8-41a315d34ec8-operator-scripts\") pod \"eb7241f8-33c3-4dde-95c8-41a315d34ec8\" (UID: \"eb7241f8-33c3-4dde-95c8-41a315d34ec8\") " Dec 05 11:06:42.009451 master-0 kubenswrapper[24928]: I1205 11:06:42.009244 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-operator-scripts\") pod \"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa\" (UID: \"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa\") " Dec 05 11:06:42.009451 master-0 kubenswrapper[24928]: I1205 11:06:42.009295 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwz2z\" (UniqueName: \"kubernetes.io/projected/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-kube-api-access-fwz2z\") pod \"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa\" (UID: \"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa\") " Dec 05 11:06:42.009451 master-0 kubenswrapper[24928]: I1205 11:06:42.009391 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plt5m\" (UniqueName: \"kubernetes.io/projected/eb7241f8-33c3-4dde-95c8-41a315d34ec8-kube-api-access-plt5m\") pod \"eb7241f8-33c3-4dde-95c8-41a315d34ec8\" (UID: \"eb7241f8-33c3-4dde-95c8-41a315d34ec8\") " Dec 05 11:06:42.010156 master-0 kubenswrapper[24928]: I1205 11:06:42.009761 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb7241f8-33c3-4dde-95c8-41a315d34ec8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eb7241f8-33c3-4dde-95c8-41a315d34ec8" (UID: "eb7241f8-33c3-4dde-95c8-41a315d34ec8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:42.010156 master-0 kubenswrapper[24928]: I1205 11:06:42.009936 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa" (UID: "55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:42.012607 master-0 kubenswrapper[24928]: I1205 11:06:42.010640 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eb7241f8-33c3-4dde-95c8-41a315d34ec8-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:42.012607 master-0 kubenswrapper[24928]: I1205 11:06:42.010668 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:42.012607 master-0 kubenswrapper[24928]: I1205 11:06:42.010663 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-conductor-0"] Dec 05 11:06:42.022990 master-0 kubenswrapper[24928]: I1205 11:06:42.015338 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb7241f8-33c3-4dde-95c8-41a315d34ec8-kube-api-access-plt5m" (OuterVolumeSpecName: "kube-api-access-plt5m") pod "eb7241f8-33c3-4dde-95c8-41a315d34ec8" (UID: "eb7241f8-33c3-4dde-95c8-41a315d34ec8"). InnerVolumeSpecName "kube-api-access-plt5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:42.022990 master-0 kubenswrapper[24928]: I1205 11:06:42.018905 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-kube-api-access-fwz2z" (OuterVolumeSpecName: "kube-api-access-fwz2z") pod "55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa" (UID: "55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa"). InnerVolumeSpecName "kube-api-access-fwz2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:42.065726 master-0 kubenswrapper[24928]: I1205 11:06:42.065630 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-ec941-volume-lvm-iscsi-0" Dec 05 11:06:42.112875 master-0 kubenswrapper[24928]: I1205 11:06:42.112326 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwz2z\" (UniqueName: \"kubernetes.io/projected/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa-kube-api-access-fwz2z\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:42.112875 master-0 kubenswrapper[24928]: I1205 11:06:42.112369 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plt5m\" (UniqueName: \"kubernetes.io/projected/eb7241f8-33c3-4dde-95c8-41a315d34ec8-kube-api-access-plt5m\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:42.561887 master-0 kubenswrapper[24928]: I1205 11:06:42.561819 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-z4xfk" event={"ID":"eb7241f8-33c3-4dde-95c8-41a315d34ec8","Type":"ContainerDied","Data":"e8a0dbd47040de38459413107627b2dac54b98ec077562a47a530a87c06abb19"} Dec 05 11:06:42.562733 master-0 kubenswrapper[24928]: I1205 11:06:42.562708 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8a0dbd47040de38459413107627b2dac54b98ec077562a47a530a87c06abb19" Dec 05 11:06:42.563064 master-0 kubenswrapper[24928]: I1205 11:06:42.562110 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-z4xfk" Dec 05 11:06:42.567964 master-0 kubenswrapper[24928]: I1205 11:06:42.567792 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-54d5ff47b5-hzncp" event={"ID":"f4540893-bc29-4b48-a741-ecb016eb0e73","Type":"ContainerStarted","Data":"fa4422211b09287950930e298e844e4ea749f555001e652c0fd0af740e913867"} Dec 05 11:06:42.569740 master-0 kubenswrapper[24928]: I1205 11:06:42.569696 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" event={"ID":"55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa","Type":"ContainerDied","Data":"6f1cd01f60de8c2e1c64b4fac2a91464250f4f60da94d08407b622a93785f63e"} Dec 05 11:06:42.569740 master-0 kubenswrapper[24928]: I1205 11:06:42.569738 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f1cd01f60de8c2e1c64b4fac2a91464250f4f60da94d08407b622a93785f63e" Dec 05 11:06:42.569949 master-0 kubenswrapper[24928]: I1205 11:06:42.569823 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-b0db-account-create-update-5x72f" Dec 05 11:06:42.571560 master-0 kubenswrapper[24928]: I1205 11:06:42.571478 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"2d8079b6-0de1-4ffd-b008-9d74bf9673ae","Type":"ContainerStarted","Data":"4f80a1bdcb107b4f0d368d9e6a79561bf7b0f74c0403caac5bd5c3d429a09511"} Dec 05 11:06:42.573674 master-0 kubenswrapper[24928]: I1205 11:06:42.573333 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-647984b79f-nk9b6" event={"ID":"1d3f7004-5e71-452f-8abd-907e10bee262","Type":"ContainerStarted","Data":"3be063f5d5616568b143ef9f671a0fa79bf310ca3a1d34a11816cd2df6467379"} Dec 05 11:06:42.580497 master-0 kubenswrapper[24928]: I1205 11:06:42.580431 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" event={"ID":"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1","Type":"ContainerStarted","Data":"e864c5cca53d872c9485934b413778043824a71ca67f0336f4c1970f63375b14"} Dec 05 11:06:42.582671 master-0 kubenswrapper[24928]: I1205 11:06:42.581932 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:42.588292 master-0 kubenswrapper[24928]: I1205 11:06:42.588222 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-ec941-backup-0" Dec 05 11:06:42.600547 master-0 kubenswrapper[24928]: I1205 11:06:42.600472 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-ec941-scheduler-0" Dec 05 11:06:43.015704 master-0 kubenswrapper[24928]: I1205 11:06:43.015441 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" podStartSLOduration=4.590322834 podStartE2EDuration="8.015407174s" podCreationTimestamp="2025-12-05 11:06:35 +0000 UTC" firstStartedPulling="2025-12-05 11:06:37.014732785 +0000 UTC m=+1157.017926626" lastFinishedPulling="2025-12-05 11:06:40.439817115 +0000 UTC m=+1160.443010966" observedRunningTime="2025-12-05 11:06:42.981191069 +0000 UTC m=+1162.984384920" watchObservedRunningTime="2025-12-05 11:06:43.015407174 +0000 UTC m=+1163.018601025" Dec 05 11:06:43.600185 master-0 kubenswrapper[24928]: I1205 11:06:43.600025 24928 generic.go:334] "Generic (PLEG): container finished" podID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerID="fa4422211b09287950930e298e844e4ea749f555001e652c0fd0af740e913867" exitCode=0 Dec 05 11:06:43.600185 master-0 kubenswrapper[24928]: I1205 11:06:43.600142 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-54d5ff47b5-hzncp" event={"ID":"f4540893-bc29-4b48-a741-ecb016eb0e73","Type":"ContainerDied","Data":"fa4422211b09287950930e298e844e4ea749f555001e652c0fd0af740e913867"} Dec 05 11:06:43.604607 master-0 kubenswrapper[24928]: I1205 11:06:43.604542 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"2d8079b6-0de1-4ffd-b008-9d74bf9673ae","Type":"ContainerStarted","Data":"505f5c68071c212fb62a19ce615c5e13c0bee7d8df8c6e9457af16927557dbe3"} Dec 05 11:06:43.606678 master-0 kubenswrapper[24928]: I1205 11:06:43.606628 24928 generic.go:334] "Generic (PLEG): container finished" podID="1d3f7004-5e71-452f-8abd-907e10bee262" containerID="f66e15de2bc7e135e0dd26f73a97962fcaebc4946357478a7a2520987474ea04" exitCode=0 Dec 05 11:06:43.606765 master-0 kubenswrapper[24928]: I1205 11:06:43.606697 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-647984b79f-nk9b6" event={"ID":"1d3f7004-5e71-452f-8abd-907e10bee262","Type":"ContainerDied","Data":"f66e15de2bc7e135e0dd26f73a97962fcaebc4946357478a7a2520987474ea04"} Dec 05 11:06:44.621494 master-0 kubenswrapper[24928]: I1205 11:06:44.621361 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-647984b79f-nk9b6" event={"ID":"1d3f7004-5e71-452f-8abd-907e10bee262","Type":"ContainerStarted","Data":"df39347d057bdf7dcfbf043cd9eb4503b76c02ac0c4d49d889055501d3220c51"} Dec 05 11:06:44.621494 master-0 kubenswrapper[24928]: I1205 11:06:44.621452 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-647984b79f-nk9b6" event={"ID":"1d3f7004-5e71-452f-8abd-907e10bee262","Type":"ContainerStarted","Data":"34ef1c2ff8dc7122cca4e2ec71d442d1d26bdaac1c867ad2beaa3c7b6fbce5df"} Dec 05 11:06:44.622405 master-0 kubenswrapper[24928]: I1205 11:06:44.622354 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:44.626191 master-0 kubenswrapper[24928]: I1205 11:06:44.626100 24928 scope.go:117] "RemoveContainer" containerID="5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212" Dec 05 11:06:44.626499 master-0 kubenswrapper[24928]: I1205 11:06:44.626464 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-54d5ff47b5-hzncp" event={"ID":"f4540893-bc29-4b48-a741-ecb016eb0e73","Type":"ContainerStarted","Data":"5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212"} Dec 05 11:06:44.626570 master-0 kubenswrapper[24928]: I1205 11:06:44.626498 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-54d5ff47b5-hzncp" event={"ID":"f4540893-bc29-4b48-a741-ecb016eb0e73","Type":"ContainerStarted","Data":"48435a82a76dfce0de134e8638b2d8aa1e1168935349387ac394b1432d91dde0"} Dec 05 11:06:44.668929 master-0 kubenswrapper[24928]: I1205 11:06:44.668855 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:44.696547 master-0 kubenswrapper[24928]: I1205 11:06:44.695879 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-647984b79f-nk9b6" podStartSLOduration=4.9870977530000005 podStartE2EDuration="5.695852389s" podCreationTimestamp="2025-12-05 11:06:39 +0000 UTC" firstStartedPulling="2025-12-05 11:06:41.789473163 +0000 UTC m=+1161.792667014" lastFinishedPulling="2025-12-05 11:06:42.498227799 +0000 UTC m=+1162.501421650" observedRunningTime="2025-12-05 11:06:44.654440828 +0000 UTC m=+1164.657634699" watchObservedRunningTime="2025-12-05 11:06:44.695852389 +0000 UTC m=+1164.699046250" Dec 05 11:06:45.287436 master-0 kubenswrapper[24928]: I1205 11:06:45.287344 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-558cb7d6b-wgm2l" Dec 05 11:06:45.646857 master-0 kubenswrapper[24928]: I1205 11:06:45.645903 24928 generic.go:334] "Generic (PLEG): container finished" podID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerID="5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212" exitCode=1 Dec 05 11:06:45.646857 master-0 kubenswrapper[24928]: I1205 11:06:45.645973 24928 generic.go:334] "Generic (PLEG): container finished" podID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerID="0bdc22ef7d01cde6d461167b49e502d552500c23ab2d5088432556d87905f364" exitCode=1 Dec 05 11:06:45.646857 master-0 kubenswrapper[24928]: I1205 11:06:45.645940 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-54d5ff47b5-hzncp" event={"ID":"f4540893-bc29-4b48-a741-ecb016eb0e73","Type":"ContainerDied","Data":"5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212"} Dec 05 11:06:45.646857 master-0 kubenswrapper[24928]: I1205 11:06:45.646039 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-54d5ff47b5-hzncp" event={"ID":"f4540893-bc29-4b48-a741-ecb016eb0e73","Type":"ContainerDied","Data":"0bdc22ef7d01cde6d461167b49e502d552500c23ab2d5088432556d87905f364"} Dec 05 11:06:45.646857 master-0 kubenswrapper[24928]: I1205 11:06:45.646059 24928 scope.go:117] "RemoveContainer" containerID="5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212" Dec 05 11:06:45.650003 master-0 kubenswrapper[24928]: I1205 11:06:45.649591 24928 generic.go:334] "Generic (PLEG): container finished" podID="2d8079b6-0de1-4ffd-b008-9d74bf9673ae" containerID="505f5c68071c212fb62a19ce615c5e13c0bee7d8df8c6e9457af16927557dbe3" exitCode=0 Dec 05 11:06:45.650003 master-0 kubenswrapper[24928]: I1205 11:06:45.649695 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"2d8079b6-0de1-4ffd-b008-9d74bf9673ae","Type":"ContainerDied","Data":"505f5c68071c212fb62a19ce615c5e13c0bee7d8df8c6e9457af16927557dbe3"} Dec 05 11:06:45.650549 master-0 kubenswrapper[24928]: I1205 11:06:45.650369 24928 scope.go:117] "RemoveContainer" containerID="0bdc22ef7d01cde6d461167b49e502d552500c23ab2d5088432556d87905f364" Dec 05 11:06:45.650883 master-0 kubenswrapper[24928]: E1205 11:06:45.650715 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-api pod=ironic-54d5ff47b5-hzncp_openstack(f4540893-bc29-4b48-a741-ecb016eb0e73)\"" pod="openstack/ironic-54d5ff47b5-hzncp" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" Dec 05 11:06:45.665587 master-0 kubenswrapper[24928]: I1205 11:06:45.665384 24928 generic.go:334] "Generic (PLEG): container finished" podID="2a9f99ee-4977-4963-a6b1-2a2bd2e849d1" containerID="e864c5cca53d872c9485934b413778043824a71ca67f0336f4c1970f63375b14" exitCode=1 Dec 05 11:06:45.665587 master-0 kubenswrapper[24928]: I1205 11:06:45.665485 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" event={"ID":"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1","Type":"ContainerDied","Data":"e864c5cca53d872c9485934b413778043824a71ca67f0336f4c1970f63375b14"} Dec 05 11:06:45.669716 master-0 kubenswrapper[24928]: I1205 11:06:45.667340 24928 scope.go:117] "RemoveContainer" containerID="e864c5cca53d872c9485934b413778043824a71ca67f0336f4c1970f63375b14" Dec 05 11:06:45.752773 master-0 kubenswrapper[24928]: I1205 11:06:45.752720 24928 scope.go:117] "RemoveContainer" containerID="5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212" Dec 05 11:06:45.753465 master-0 kubenswrapper[24928]: E1205 11:06:45.753375 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212\": container with ID starting with 5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212 not found: ID does not exist" containerID="5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212" Dec 05 11:06:45.753465 master-0 kubenswrapper[24928]: I1205 11:06:45.753454 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212"} err="failed to get container status \"5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212\": rpc error: code = NotFound desc = could not find container \"5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212\": container with ID starting with 5ed9c4ee3ac7f5899e8c5da650eaa6d8945a8a9d23a9232e039650fe3da4e212 not found: ID does not exist" Dec 05 11:06:46.391767 master-0 kubenswrapper[24928]: I1205 11:06:46.391692 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:46.391767 master-0 kubenswrapper[24928]: I1205 11:06:46.391743 24928 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:46.673795 master-0 kubenswrapper[24928]: I1205 11:06:46.673595 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:06:46.688925 master-0 kubenswrapper[24928]: I1205 11:06:46.688850 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:46.688925 master-0 kubenswrapper[24928]: I1205 11:06:46.688914 24928 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:46.689878 master-0 kubenswrapper[24928]: I1205 11:06:46.689797 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" event={"ID":"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1","Type":"ContainerStarted","Data":"1b0154660cc27dc1ba8da7a2dc97051ffd00f6f4aa5da15396421e670962c1d7"} Dec 05 11:06:46.690061 master-0 kubenswrapper[24928]: I1205 11:06:46.690026 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:46.694181 master-0 kubenswrapper[24928]: I1205 11:06:46.694139 24928 scope.go:117] "RemoveContainer" containerID="0bdc22ef7d01cde6d461167b49e502d552500c23ab2d5088432556d87905f364" Dec 05 11:06:46.694824 master-0 kubenswrapper[24928]: E1205 11:06:46.694504 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-api pod=ironic-54d5ff47b5-hzncp_openstack(f4540893-bc29-4b48-a741-ecb016eb0e73)\"" pod="openstack/ironic-54d5ff47b5-hzncp" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" Dec 05 11:06:47.127540 master-0 kubenswrapper[24928]: I1205 11:06:47.119624 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59b84c6dcc-54hsm"] Dec 05 11:06:47.127540 master-0 kubenswrapper[24928]: I1205 11:06:47.120012 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" podUID="c684b0cd-620b-45f3-9fee-4eb6f8cf99df" containerName="dnsmasq-dns" containerID="cri-o://37c7fa0465c346fadab41fa46614e8152d9d6def2ed2ac30169efb28ed50a342" gracePeriod=10 Dec 05 11:06:47.715155 master-0 kubenswrapper[24928]: I1205 11:06:47.715013 24928 generic.go:334] "Generic (PLEG): container finished" podID="c684b0cd-620b-45f3-9fee-4eb6f8cf99df" containerID="37c7fa0465c346fadab41fa46614e8152d9d6def2ed2ac30169efb28ed50a342" exitCode=0 Dec 05 11:06:47.715768 master-0 kubenswrapper[24928]: I1205 11:06:47.715119 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" event={"ID":"c684b0cd-620b-45f3-9fee-4eb6f8cf99df","Type":"ContainerDied","Data":"37c7fa0465c346fadab41fa46614e8152d9d6def2ed2ac30169efb28ed50a342"} Dec 05 11:06:47.716349 master-0 kubenswrapper[24928]: I1205 11:06:47.716305 24928 scope.go:117] "RemoveContainer" containerID="0bdc22ef7d01cde6d461167b49e502d552500c23ab2d5088432556d87905f364" Dec 05 11:06:47.716753 master-0 kubenswrapper[24928]: E1205 11:06:47.716721 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-api pod=ironic-54d5ff47b5-hzncp_openstack(f4540893-bc29-4b48-a741-ecb016eb0e73)\"" pod="openstack/ironic-54d5ff47b5-hzncp" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" Dec 05 11:06:48.688446 master-0 kubenswrapper[24928]: I1205 11:06:48.688315 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:48.699332 master-0 kubenswrapper[24928]: I1205 11:06:48.699262 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:06:48.758925 master-0 kubenswrapper[24928]: I1205 11:06:48.758849 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" event={"ID":"c684b0cd-620b-45f3-9fee-4eb6f8cf99df","Type":"ContainerDied","Data":"c2d9e6b2ccef35e52011744ae9a644a93ff1041922f84f34e5e7840bd1d9afa2"} Dec 05 11:06:48.758925 master-0 kubenswrapper[24928]: I1205 11:06:48.758923 24928 scope.go:117] "RemoveContainer" containerID="37c7fa0465c346fadab41fa46614e8152d9d6def2ed2ac30169efb28ed50a342" Dec 05 11:06:48.759539 master-0 kubenswrapper[24928]: I1205 11:06:48.758921 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59b84c6dcc-54hsm" Dec 05 11:06:48.794369 master-0 kubenswrapper[24928]: I1205 11:06:48.794328 24928 scope.go:117] "RemoveContainer" containerID="b259bd55744f87238af7ef1659a02b7493e7a2fd6bf68f40b00bbea8b402739a" Dec 05 11:06:48.875996 master-0 kubenswrapper[24928]: I1205 11:06:48.875926 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-config\") pod \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " Dec 05 11:06:48.876560 master-0 kubenswrapper[24928]: I1205 11:06:48.876021 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpmht\" (UniqueName: \"kubernetes.io/projected/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-kube-api-access-gpmht\") pod \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " Dec 05 11:06:48.876635 master-0 kubenswrapper[24928]: I1205 11:06:48.876601 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-sb\") pod \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " Dec 05 11:06:48.877305 master-0 kubenswrapper[24928]: I1205 11:06:48.877032 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-svc\") pod \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " Dec 05 11:06:48.877491 master-0 kubenswrapper[24928]: I1205 11:06:48.877236 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-nb\") pod \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " Dec 05 11:06:48.877594 master-0 kubenswrapper[24928]: I1205 11:06:48.877568 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-swift-storage-0\") pod \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\" (UID: \"c684b0cd-620b-45f3-9fee-4eb6f8cf99df\") " Dec 05 11:06:48.885925 master-0 kubenswrapper[24928]: I1205 11:06:48.885827 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-kube-api-access-gpmht" (OuterVolumeSpecName: "kube-api-access-gpmht") pod "c684b0cd-620b-45f3-9fee-4eb6f8cf99df" (UID: "c684b0cd-620b-45f3-9fee-4eb6f8cf99df"). InnerVolumeSpecName "kube-api-access-gpmht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:48.943906 master-0 kubenswrapper[24928]: I1205 11:06:48.943623 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c684b0cd-620b-45f3-9fee-4eb6f8cf99df" (UID: "c684b0cd-620b-45f3-9fee-4eb6f8cf99df"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:48.946180 master-0 kubenswrapper[24928]: I1205 11:06:48.946098 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c684b0cd-620b-45f3-9fee-4eb6f8cf99df" (UID: "c684b0cd-620b-45f3-9fee-4eb6f8cf99df"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:48.948315 master-0 kubenswrapper[24928]: I1205 11:06:48.948220 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c684b0cd-620b-45f3-9fee-4eb6f8cf99df" (UID: "c684b0cd-620b-45f3-9fee-4eb6f8cf99df"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:48.953958 master-0 kubenswrapper[24928]: I1205 11:06:48.953899 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-config" (OuterVolumeSpecName: "config") pod "c684b0cd-620b-45f3-9fee-4eb6f8cf99df" (UID: "c684b0cd-620b-45f3-9fee-4eb6f8cf99df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:48.954092 master-0 kubenswrapper[24928]: I1205 11:06:48.953999 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c684b0cd-620b-45f3-9fee-4eb6f8cf99df" (UID: "c684b0cd-620b-45f3-9fee-4eb6f8cf99df"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:06:48.981644 master-0 kubenswrapper[24928]: I1205 11:06:48.981509 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:48.981644 master-0 kubenswrapper[24928]: I1205 11:06:48.981565 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:48.981644 master-0 kubenswrapper[24928]: I1205 11:06:48.981582 24928 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-dns-swift-storage-0\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:48.981644 master-0 kubenswrapper[24928]: I1205 11:06:48.981594 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:48.981644 master-0 kubenswrapper[24928]: I1205 11:06:48.981607 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpmht\" (UniqueName: \"kubernetes.io/projected/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-kube-api-access-gpmht\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:48.981644 master-0 kubenswrapper[24928]: I1205 11:06:48.981619 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c684b0cd-620b-45f3-9fee-4eb6f8cf99df-ovsdbserver-sb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:49.261467 master-0 kubenswrapper[24928]: I1205 11:06:49.255152 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59b84c6dcc-54hsm"] Dec 05 11:06:49.288934 master-0 kubenswrapper[24928]: I1205 11:06:49.288863 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59b84c6dcc-54hsm"] Dec 05 11:06:49.783335 master-0 kubenswrapper[24928]: I1205 11:06:49.783169 24928 generic.go:334] "Generic (PLEG): container finished" podID="2a9f99ee-4977-4963-a6b1-2a2bd2e849d1" containerID="1b0154660cc27dc1ba8da7a2dc97051ffd00f6f4aa5da15396421e670962c1d7" exitCode=1 Dec 05 11:06:49.783335 master-0 kubenswrapper[24928]: I1205 11:06:49.783230 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" event={"ID":"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1","Type":"ContainerDied","Data":"1b0154660cc27dc1ba8da7a2dc97051ffd00f6f4aa5da15396421e670962c1d7"} Dec 05 11:06:49.783335 master-0 kubenswrapper[24928]: I1205 11:06:49.783268 24928 scope.go:117] "RemoveContainer" containerID="e864c5cca53d872c9485934b413778043824a71ca67f0336f4c1970f63375b14" Dec 05 11:06:49.785238 master-0 kubenswrapper[24928]: I1205 11:06:49.785074 24928 scope.go:117] "RemoveContainer" containerID="1b0154660cc27dc1ba8da7a2dc97051ffd00f6f4aa5da15396421e670962c1d7" Dec 05 11:06:49.787809 master-0 kubenswrapper[24928]: E1205 11:06:49.787668 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-66b77f87b6-l2d4s_openstack(2a9f99ee-4977-4963-a6b1-2a2bd2e849d1)\"" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" podUID="2a9f99ee-4977-4963-a6b1-2a2bd2e849d1" Dec 05 11:06:50.254722 master-0 kubenswrapper[24928]: I1205 11:06:50.228382 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c684b0cd-620b-45f3-9fee-4eb6f8cf99df" path="/var/lib/kubelet/pods/c684b0cd-620b-45f3-9fee-4eb6f8cf99df/volumes" Dec 05 11:06:50.254935 master-0 kubenswrapper[24928]: I1205 11:06:50.254733 24928 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","poda9d6834d-6213-4cdc-9f0b-b336e4cf2ee3"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort poda9d6834d-6213-4cdc-9f0b-b336e4cf2ee3] : Timed out while waiting for systemd to remove kubepods-besteffort-poda9d6834d_6213_4cdc_9f0b_b336e4cf2ee3.slice" Dec 05 11:06:50.254935 master-0 kubenswrapper[24928]: E1205 11:06:50.254817 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort poda9d6834d-6213-4cdc-9f0b-b336e4cf2ee3] : unable to destroy cgroup paths for cgroup [kubepods besteffort poda9d6834d-6213-4cdc-9f0b-b336e4cf2ee3] : Timed out while waiting for systemd to remove kubepods-besteffort-poda9d6834d_6213_4cdc_9f0b_b336e4cf2ee3.slice" pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" podUID="a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" Dec 05 11:06:50.794680 master-0 kubenswrapper[24928]: I1205 11:06:50.794626 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-559ff6cd8f-5sc8w" Dec 05 11:06:51.391234 master-0 kubenswrapper[24928]: I1205 11:06:51.391167 24928 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:06:51.392114 master-0 kubenswrapper[24928]: I1205 11:06:51.392084 24928 scope.go:117] "RemoveContainer" containerID="1b0154660cc27dc1ba8da7a2dc97051ffd00f6f4aa5da15396421e670962c1d7" Dec 05 11:06:51.392399 master-0 kubenswrapper[24928]: E1205 11:06:51.392371 24928 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-66b77f87b6-l2d4s_openstack(2a9f99ee-4977-4963-a6b1-2a2bd2e849d1)\"" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" podUID="2a9f99ee-4977-4963-a6b1-2a2bd2e849d1" Dec 05 11:06:52.692011 master-0 kubenswrapper[24928]: I1205 11:06:52.691919 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 11:06:52.692682 master-0 kubenswrapper[24928]: E1205 11:06:52.692574 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb7241f8-33c3-4dde-95c8-41a315d34ec8" containerName="mariadb-database-create" Dec 05 11:06:52.692682 master-0 kubenswrapper[24928]: I1205 11:06:52.692591 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb7241f8-33c3-4dde-95c8-41a315d34ec8" containerName="mariadb-database-create" Dec 05 11:06:52.692682 master-0 kubenswrapper[24928]: E1205 11:06:52.692604 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c684b0cd-620b-45f3-9fee-4eb6f8cf99df" containerName="init" Dec 05 11:06:52.692682 master-0 kubenswrapper[24928]: I1205 11:06:52.692609 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c684b0cd-620b-45f3-9fee-4eb6f8cf99df" containerName="init" Dec 05 11:06:52.692682 master-0 kubenswrapper[24928]: E1205 11:06:52.692622 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c684b0cd-620b-45f3-9fee-4eb6f8cf99df" containerName="dnsmasq-dns" Dec 05 11:06:52.692682 master-0 kubenswrapper[24928]: I1205 11:06:52.692629 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="c684b0cd-620b-45f3-9fee-4eb6f8cf99df" containerName="dnsmasq-dns" Dec 05 11:06:52.692682 master-0 kubenswrapper[24928]: E1205 11:06:52.692644 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa" containerName="mariadb-account-create-update" Dec 05 11:06:52.692682 master-0 kubenswrapper[24928]: I1205 11:06:52.692652 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa" containerName="mariadb-account-create-update" Dec 05 11:06:52.693067 master-0 kubenswrapper[24928]: I1205 11:06:52.692885 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb7241f8-33c3-4dde-95c8-41a315d34ec8" containerName="mariadb-database-create" Dec 05 11:06:52.693067 master-0 kubenswrapper[24928]: I1205 11:06:52.692906 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="c684b0cd-620b-45f3-9fee-4eb6f8cf99df" containerName="dnsmasq-dns" Dec 05 11:06:52.693067 master-0 kubenswrapper[24928]: I1205 11:06:52.692927 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa" containerName="mariadb-account-create-update" Dec 05 11:06:52.693769 master-0 kubenswrapper[24928]: I1205 11:06:52.693734 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 11:06:52.696366 master-0 kubenswrapper[24928]: I1205 11:06:52.696306 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 05 11:06:52.697785 master-0 kubenswrapper[24928]: I1205 11:06:52.697317 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 05 11:06:52.762993 master-0 kubenswrapper[24928]: I1205 11:06:52.762215 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 11:06:52.766229 master-0 kubenswrapper[24928]: I1205 11:06:52.766057 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-647984b79f-nk9b6" Dec 05 11:06:52.790485 master-0 kubenswrapper[24928]: I1205 11:06:52.790383 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fa285890-6077-4460-8935-e683d1c36783-openstack-config\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:52.790638 master-0 kubenswrapper[24928]: I1205 11:06:52.790597 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa285890-6077-4460-8935-e683d1c36783-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:52.790698 master-0 kubenswrapper[24928]: I1205 11:06:52.790679 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56cd9\" (UniqueName: \"kubernetes.io/projected/fa285890-6077-4460-8935-e683d1c36783-kube-api-access-56cd9\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:52.790755 master-0 kubenswrapper[24928]: I1205 11:06:52.790736 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fa285890-6077-4460-8935-e683d1c36783-openstack-config-secret\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:52.882493 master-0 kubenswrapper[24928]: I1205 11:06:52.881487 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-559ff6cd8f-5sc8w"] Dec 05 11:06:52.893065 master-0 kubenswrapper[24928]: I1205 11:06:52.893002 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa285890-6077-4460-8935-e683d1c36783-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:52.893348 master-0 kubenswrapper[24928]: I1205 11:06:52.893136 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56cd9\" (UniqueName: \"kubernetes.io/projected/fa285890-6077-4460-8935-e683d1c36783-kube-api-access-56cd9\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:52.893348 master-0 kubenswrapper[24928]: I1205 11:06:52.893234 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fa285890-6077-4460-8935-e683d1c36783-openstack-config-secret\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:52.893348 master-0 kubenswrapper[24928]: I1205 11:06:52.893324 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fa285890-6077-4460-8935-e683d1c36783-openstack-config\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:52.896062 master-0 kubenswrapper[24928]: I1205 11:06:52.896008 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-559ff6cd8f-5sc8w"] Dec 05 11:06:52.897146 master-0 kubenswrapper[24928]: I1205 11:06:52.897088 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fa285890-6077-4460-8935-e683d1c36783-combined-ca-bundle\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:52.899308 master-0 kubenswrapper[24928]: I1205 11:06:52.899102 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/fa285890-6077-4460-8935-e683d1c36783-openstack-config\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:52.924559 master-0 kubenswrapper[24928]: I1205 11:06:52.917471 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56cd9\" (UniqueName: \"kubernetes.io/projected/fa285890-6077-4460-8935-e683d1c36783-kube-api-access-56cd9\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:52.924559 master-0 kubenswrapper[24928]: I1205 11:06:52.921699 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/fa285890-6077-4460-8935-e683d1c36783-openstack-config-secret\") pod \"openstackclient\" (UID: \"fa285890-6077-4460-8935-e683d1c36783\") " pod="openstack/openstackclient" Dec 05 11:06:53.013280 master-0 kubenswrapper[24928]: I1205 11:06:53.013199 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 11:06:53.137882 master-0 kubenswrapper[24928]: I1205 11:06:53.137735 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-54d5ff47b5-hzncp"] Dec 05 11:06:53.138150 master-0 kubenswrapper[24928]: I1205 11:06:53.138090 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ironic-54d5ff47b5-hzncp" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="ironic-api-log" containerID="cri-o://48435a82a76dfce0de134e8638b2d8aa1e1168935349387ac394b1432d91dde0" gracePeriod=60 Dec 05 11:06:53.148202 master-0 kubenswrapper[24928]: I1205 11:06:53.148112 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-db-sync-sppfx"] Dec 05 11:06:53.150650 master-0 kubenswrapper[24928]: I1205 11:06:53.150621 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.153432 master-0 kubenswrapper[24928]: I1205 11:06:53.153389 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Dec 05 11:06:53.153564 master-0 kubenswrapper[24928]: I1205 11:06:53.153519 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Dec 05 11:06:53.158563 master-0 kubenswrapper[24928]: I1205 11:06:53.158507 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-sync-sppfx"] Dec 05 11:06:53.319492 master-0 kubenswrapper[24928]: I1205 11:06:53.319345 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.319709 master-0 kubenswrapper[24928]: I1205 11:06:53.319562 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-scripts\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.319709 master-0 kubenswrapper[24928]: I1205 11:06:53.319630 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5183387d-f79b-4412-88ac-41b37866e07e-etc-podinfo\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.319709 master-0 kubenswrapper[24928]: I1205 11:06:53.319657 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-config\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.319824 master-0 kubenswrapper[24928]: I1205 11:06:53.319718 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-combined-ca-bundle\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.319824 master-0 kubenswrapper[24928]: I1205 11:06:53.319778 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.319883 master-0 kubenswrapper[24928]: I1205 11:06:53.319842 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb5w8\" (UniqueName: \"kubernetes.io/projected/5183387d-f79b-4412-88ac-41b37866e07e-kube-api-access-wb5w8\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.422990 master-0 kubenswrapper[24928]: I1205 11:06:53.422914 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-scripts\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.423253 master-0 kubenswrapper[24928]: I1205 11:06:53.423107 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5183387d-f79b-4412-88ac-41b37866e07e-etc-podinfo\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.423253 master-0 kubenswrapper[24928]: I1205 11:06:53.423151 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-config\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.423914 master-0 kubenswrapper[24928]: I1205 11:06:53.423410 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-combined-ca-bundle\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.424048 master-0 kubenswrapper[24928]: I1205 11:06:53.424017 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.424214 master-0 kubenswrapper[24928]: I1205 11:06:53.424177 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb5w8\" (UniqueName: \"kubernetes.io/projected/5183387d-f79b-4412-88ac-41b37866e07e-kube-api-access-wb5w8\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.424606 master-0 kubenswrapper[24928]: I1205 11:06:53.424555 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.424801 master-0 kubenswrapper[24928]: I1205 11:06:53.424575 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.424941 master-0 kubenswrapper[24928]: I1205 11:06:53.424910 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.427044 master-0 kubenswrapper[24928]: I1205 11:06:53.426967 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5183387d-f79b-4412-88ac-41b37866e07e-etc-podinfo\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.427936 master-0 kubenswrapper[24928]: I1205 11:06:53.427892 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-combined-ca-bundle\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.428358 master-0 kubenswrapper[24928]: I1205 11:06:53.428315 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-config\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.430202 master-0 kubenswrapper[24928]: I1205 11:06:53.430160 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-scripts\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.490402 master-0 kubenswrapper[24928]: I1205 11:06:53.490352 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb5w8\" (UniqueName: \"kubernetes.io/projected/5183387d-f79b-4412-88ac-41b37866e07e-kube-api-access-wb5w8\") pod \"ironic-inspector-db-sync-sppfx\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.561907 master-0 kubenswrapper[24928]: I1205 11:06:53.561852 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:06:53.878439 master-0 kubenswrapper[24928]: I1205 11:06:53.873649 24928 generic.go:334] "Generic (PLEG): container finished" podID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerID="48435a82a76dfce0de134e8638b2d8aa1e1168935349387ac394b1432d91dde0" exitCode=143 Dec 05 11:06:53.878439 master-0 kubenswrapper[24928]: I1205 11:06:53.873712 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-54d5ff47b5-hzncp" event={"ID":"f4540893-bc29-4b48-a741-ecb016eb0e73","Type":"ContainerDied","Data":"48435a82a76dfce0de134e8638b2d8aa1e1168935349387ac394b1432d91dde0"} Dec 05 11:06:53.878439 master-0 kubenswrapper[24928]: I1205 11:06:53.873788 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-54d5ff47b5-hzncp" event={"ID":"f4540893-bc29-4b48-a741-ecb016eb0e73","Type":"ContainerDied","Data":"40de73483f3d37ce317b8d6af0082084388fd8d7a2b1ae8a16b5008f83dc874f"} Dec 05 11:06:53.878439 master-0 kubenswrapper[24928]: I1205 11:06:53.873807 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40de73483f3d37ce317b8d6af0082084388fd8d7a2b1ae8a16b5008f83dc874f" Dec 05 11:06:53.887336 master-0 kubenswrapper[24928]: W1205 11:06:53.887252 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa285890_6077_4460_8935_e683d1c36783.slice/crio-9e6935b3a4ea9257c40c4ce08561c4f5fdb805dbf5f1e73a84f477ba727d5223 WatchSource:0}: Error finding container 9e6935b3a4ea9257c40c4ce08561c4f5fdb805dbf5f1e73a84f477ba727d5223: Status 404 returned error can't find the container with id 9e6935b3a4ea9257c40c4ce08561c4f5fdb805dbf5f1e73a84f477ba727d5223 Dec 05 11:06:53.920839 master-0 kubenswrapper[24928]: I1205 11:06:53.920768 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 11:06:53.962948 master-0 kubenswrapper[24928]: I1205 11:06:53.962897 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:54.039466 master-0 kubenswrapper[24928]: I1205 11:06:54.038127 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-scripts\") pod \"f4540893-bc29-4b48-a741-ecb016eb0e73\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " Dec 05 11:06:54.039466 master-0 kubenswrapper[24928]: I1205 11:06:54.038276 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-merged\") pod \"f4540893-bc29-4b48-a741-ecb016eb0e73\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " Dec 05 11:06:54.039466 master-0 kubenswrapper[24928]: I1205 11:06:54.038308 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-logs\") pod \"f4540893-bc29-4b48-a741-ecb016eb0e73\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " Dec 05 11:06:54.039466 master-0 kubenswrapper[24928]: I1205 11:06:54.038435 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f4540893-bc29-4b48-a741-ecb016eb0e73-etc-podinfo\") pod \"f4540893-bc29-4b48-a741-ecb016eb0e73\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " Dec 05 11:06:54.039466 master-0 kubenswrapper[24928]: I1205 11:06:54.038466 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-combined-ca-bundle\") pod \"f4540893-bc29-4b48-a741-ecb016eb0e73\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " Dec 05 11:06:54.039466 master-0 kubenswrapper[24928]: I1205 11:06:54.038519 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data\") pod \"f4540893-bc29-4b48-a741-ecb016eb0e73\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " Dec 05 11:06:54.039466 master-0 kubenswrapper[24928]: I1205 11:06:54.038545 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-custom\") pod \"f4540893-bc29-4b48-a741-ecb016eb0e73\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " Dec 05 11:06:54.039466 master-0 kubenswrapper[24928]: I1205 11:06:54.038583 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7xts\" (UniqueName: \"kubernetes.io/projected/f4540893-bc29-4b48-a741-ecb016eb0e73-kube-api-access-s7xts\") pod \"f4540893-bc29-4b48-a741-ecb016eb0e73\" (UID: \"f4540893-bc29-4b48-a741-ecb016eb0e73\") " Dec 05 11:06:54.040482 master-0 kubenswrapper[24928]: I1205 11:06:54.040319 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-logs" (OuterVolumeSpecName: "logs") pod "f4540893-bc29-4b48-a741-ecb016eb0e73" (UID: "f4540893-bc29-4b48-a741-ecb016eb0e73"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:06:54.040714 master-0 kubenswrapper[24928]: I1205 11:06:54.040631 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "f4540893-bc29-4b48-a741-ecb016eb0e73" (UID: "f4540893-bc29-4b48-a741-ecb016eb0e73"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:06:54.041624 master-0 kubenswrapper[24928]: I1205 11:06:54.041579 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/f4540893-bc29-4b48-a741-ecb016eb0e73-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "f4540893-bc29-4b48-a741-ecb016eb0e73" (UID: "f4540893-bc29-4b48-a741-ecb016eb0e73"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 11:06:54.042265 master-0 kubenswrapper[24928]: I1205 11:06:54.042207 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-scripts" (OuterVolumeSpecName: "scripts") pod "f4540893-bc29-4b48-a741-ecb016eb0e73" (UID: "f4540893-bc29-4b48-a741-ecb016eb0e73"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:54.044522 master-0 kubenswrapper[24928]: I1205 11:06:54.044445 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f4540893-bc29-4b48-a741-ecb016eb0e73" (UID: "f4540893-bc29-4b48-a741-ecb016eb0e73"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:54.053187 master-0 kubenswrapper[24928]: I1205 11:06:54.052318 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4540893-bc29-4b48-a741-ecb016eb0e73-kube-api-access-s7xts" (OuterVolumeSpecName: "kube-api-access-s7xts") pod "f4540893-bc29-4b48-a741-ecb016eb0e73" (UID: "f4540893-bc29-4b48-a741-ecb016eb0e73"). InnerVolumeSpecName "kube-api-access-s7xts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:06:54.109113 master-0 kubenswrapper[24928]: I1205 11:06:54.108650 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data" (OuterVolumeSpecName: "config-data") pod "f4540893-bc29-4b48-a741-ecb016eb0e73" (UID: "f4540893-bc29-4b48-a741-ecb016eb0e73"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:54.127152 master-0 kubenswrapper[24928]: I1205 11:06:54.127099 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4540893-bc29-4b48-a741-ecb016eb0e73" (UID: "f4540893-bc29-4b48-a741-ecb016eb0e73"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:06:54.129012 master-0 kubenswrapper[24928]: I1205 11:06:54.128978 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7c965f7db6-5tq2s"] Dec 05 11:06:54.129663 master-0 kubenswrapper[24928]: E1205 11:06:54.129639 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="ironic-api" Dec 05 11:06:54.129663 master-0 kubenswrapper[24928]: I1205 11:06:54.129661 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="ironic-api" Dec 05 11:06:54.129765 master-0 kubenswrapper[24928]: E1205 11:06:54.129688 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="ironic-api" Dec 05 11:06:54.129765 master-0 kubenswrapper[24928]: I1205 11:06:54.129695 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="ironic-api" Dec 05 11:06:54.129765 master-0 kubenswrapper[24928]: E1205 11:06:54.129709 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="init" Dec 05 11:06:54.129765 master-0 kubenswrapper[24928]: I1205 11:06:54.129719 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="init" Dec 05 11:06:54.129884 master-0 kubenswrapper[24928]: E1205 11:06:54.129768 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="ironic-api-log" Dec 05 11:06:54.129884 master-0 kubenswrapper[24928]: I1205 11:06:54.129776 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="ironic-api-log" Dec 05 11:06:54.130022 master-0 kubenswrapper[24928]: I1205 11:06:54.130002 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="ironic-api" Dec 05 11:06:54.130069 master-0 kubenswrapper[24928]: I1205 11:06:54.130030 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="ironic-api-log" Dec 05 11:06:54.130069 master-0 kubenswrapper[24928]: I1205 11:06:54.130057 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" containerName="ironic-api" Dec 05 11:06:54.131311 master-0 kubenswrapper[24928]: I1205 11:06:54.131222 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.137589 master-0 kubenswrapper[24928]: W1205 11:06:54.137470 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5183387d_f79b_4412_88ac_41b37866e07e.slice/crio-8a2ffae54ef8df93a24034a3fee4600ec48902cf8194e530070602b66a2f25aa WatchSource:0}: Error finding container 8a2ffae54ef8df93a24034a3fee4600ec48902cf8194e530070602b66a2f25aa: Status 404 returned error can't find the container with id 8a2ffae54ef8df93a24034a3fee4600ec48902cf8194e530070602b66a2f25aa Dec 05 11:06:54.138146 master-0 kubenswrapper[24928]: I1205 11:06:54.137760 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 05 11:06:54.138146 master-0 kubenswrapper[24928]: I1205 11:06:54.137958 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 05 11:06:54.138146 master-0 kubenswrapper[24928]: I1205 11:06:54.138065 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 05 11:06:54.143376 master-0 kubenswrapper[24928]: I1205 11:06:54.141929 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:54.143376 master-0 kubenswrapper[24928]: I1205 11:06:54.141974 24928 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-custom\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:54.143376 master-0 kubenswrapper[24928]: I1205 11:06:54.141988 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7xts\" (UniqueName: \"kubernetes.io/projected/f4540893-bc29-4b48-a741-ecb016eb0e73-kube-api-access-s7xts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:54.143376 master-0 kubenswrapper[24928]: I1205 11:06:54.142000 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:54.143376 master-0 kubenswrapper[24928]: I1205 11:06:54.142008 24928 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-config-data-merged\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:54.143376 master-0 kubenswrapper[24928]: I1205 11:06:54.142016 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4540893-bc29-4b48-a741-ecb016eb0e73-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:54.143376 master-0 kubenswrapper[24928]: I1205 11:06:54.142025 24928 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f4540893-bc29-4b48-a741-ecb016eb0e73-etc-podinfo\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:54.143376 master-0 kubenswrapper[24928]: I1205 11:06:54.142033 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4540893-bc29-4b48-a741-ecb016eb0e73-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:06:54.194732 master-0 kubenswrapper[24928]: I1205 11:06:54.194650 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-sync-sppfx"] Dec 05 11:06:54.196667 master-0 kubenswrapper[24928]: I1205 11:06:54.196614 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7c95d944d9-82sjq" Dec 05 11:06:54.223278 master-0 kubenswrapper[24928]: I1205 11:06:54.222712 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3" path="/var/lib/kubelet/pods/a9d6834d-6213-4cdc-9f0b-b336e4cf2ee3/volumes" Dec 05 11:06:54.223708 master-0 kubenswrapper[24928]: I1205 11:06:54.223609 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7c965f7db6-5tq2s"] Dec 05 11:06:54.244177 master-0 kubenswrapper[24928]: I1205 11:06:54.244097 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r48rr\" (UniqueName: \"kubernetes.io/projected/480d9df9-54a7-457d-8785-0c6074ea33f7-kube-api-access-r48rr\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.244347 master-0 kubenswrapper[24928]: I1205 11:06:54.244213 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/480d9df9-54a7-457d-8785-0c6074ea33f7-etc-swift\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.244347 master-0 kubenswrapper[24928]: I1205 11:06:54.244278 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-combined-ca-bundle\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.244687 master-0 kubenswrapper[24928]: I1205 11:06:54.244367 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-config-data\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.244687 master-0 kubenswrapper[24928]: I1205 11:06:54.244597 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/480d9df9-54a7-457d-8785-0c6074ea33f7-log-httpd\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.244687 master-0 kubenswrapper[24928]: I1205 11:06:54.244665 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-public-tls-certs\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.244836 master-0 kubenswrapper[24928]: I1205 11:06:54.244745 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/480d9df9-54a7-457d-8785-0c6074ea33f7-run-httpd\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.244889 master-0 kubenswrapper[24928]: I1205 11:06:54.244861 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-internal-tls-certs\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.347197 master-0 kubenswrapper[24928]: I1205 11:06:54.346921 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-combined-ca-bundle\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.347197 master-0 kubenswrapper[24928]: I1205 11:06:54.347006 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-config-data\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.347480 master-0 kubenswrapper[24928]: I1205 11:06:54.347251 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/480d9df9-54a7-457d-8785-0c6074ea33f7-log-httpd\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.347480 master-0 kubenswrapper[24928]: I1205 11:06:54.347337 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-public-tls-certs\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.347480 master-0 kubenswrapper[24928]: I1205 11:06:54.347398 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/480d9df9-54a7-457d-8785-0c6074ea33f7-run-httpd\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.347693 master-0 kubenswrapper[24928]: I1205 11:06:54.347560 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-internal-tls-certs\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.348124 master-0 kubenswrapper[24928]: I1205 11:06:54.347875 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/480d9df9-54a7-457d-8785-0c6074ea33f7-log-httpd\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.348653 master-0 kubenswrapper[24928]: I1205 11:06:54.348553 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r48rr\" (UniqueName: \"kubernetes.io/projected/480d9df9-54a7-457d-8785-0c6074ea33f7-kube-api-access-r48rr\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.348743 master-0 kubenswrapper[24928]: I1205 11:06:54.348725 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/480d9df9-54a7-457d-8785-0c6074ea33f7-run-httpd\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.349291 master-0 kubenswrapper[24928]: I1205 11:06:54.349251 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/480d9df9-54a7-457d-8785-0c6074ea33f7-etc-swift\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.353219 master-0 kubenswrapper[24928]: I1205 11:06:54.353142 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-combined-ca-bundle\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.353960 master-0 kubenswrapper[24928]: I1205 11:06:54.353916 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-config-data\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.354444 master-0 kubenswrapper[24928]: I1205 11:06:54.354102 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-internal-tls-certs\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.357376 master-0 kubenswrapper[24928]: I1205 11:06:54.354819 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/480d9df9-54a7-457d-8785-0c6074ea33f7-etc-swift\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.357376 master-0 kubenswrapper[24928]: I1205 11:06:54.355494 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/480d9df9-54a7-457d-8785-0c6074ea33f7-public-tls-certs\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.580843 master-0 kubenswrapper[24928]: I1205 11:06:54.578802 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r48rr\" (UniqueName: \"kubernetes.io/projected/480d9df9-54a7-457d-8785-0c6074ea33f7-kube-api-access-r48rr\") pod \"swift-proxy-7c965f7db6-5tq2s\" (UID: \"480d9df9-54a7-457d-8785-0c6074ea33f7\") " pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.619705 master-0 kubenswrapper[24928]: I1205 11:06:54.619603 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bd67d49b4-fmw2f"] Dec 05 11:06:54.619985 master-0 kubenswrapper[24928]: I1205 11:06:54.619922 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bd67d49b4-fmw2f" podUID="81360b3f-3235-4cc7-9ef3-e8f748037762" containerName="neutron-api" containerID="cri-o://66554b0348b461b3687234b1ec9f9762acaa91069b3301f211e96317a7da9a8e" gracePeriod=30 Dec 05 11:06:54.620307 master-0 kubenswrapper[24928]: I1205 11:06:54.620263 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7bd67d49b4-fmw2f" podUID="81360b3f-3235-4cc7-9ef3-e8f748037762" containerName="neutron-httpd" containerID="cri-o://e086a3c64a2da4fb8778731203db97137f5bdf9bf355be905b73134e8fd137d3" gracePeriod=30 Dec 05 11:06:54.770201 master-0 kubenswrapper[24928]: I1205 11:06:54.769997 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:54.898575 master-0 kubenswrapper[24928]: I1205 11:06:54.892576 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"fa285890-6077-4460-8935-e683d1c36783","Type":"ContainerStarted","Data":"9e6935b3a4ea9257c40c4ce08561c4f5fdb805dbf5f1e73a84f477ba727d5223"} Dec 05 11:06:54.898575 master-0 kubenswrapper[24928]: I1205 11:06:54.894925 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-sppfx" event={"ID":"5183387d-f79b-4412-88ac-41b37866e07e","Type":"ContainerStarted","Data":"8a2ffae54ef8df93a24034a3fee4600ec48902cf8194e530070602b66a2f25aa"} Dec 05 11:06:54.916712 master-0 kubenswrapper[24928]: I1205 11:06:54.900683 24928 generic.go:334] "Generic (PLEG): container finished" podID="81360b3f-3235-4cc7-9ef3-e8f748037762" containerID="e086a3c64a2da4fb8778731203db97137f5bdf9bf355be905b73134e8fd137d3" exitCode=0 Dec 05 11:06:54.916712 master-0 kubenswrapper[24928]: I1205 11:06:54.900769 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-54d5ff47b5-hzncp" Dec 05 11:06:54.916712 master-0 kubenswrapper[24928]: I1205 11:06:54.901801 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd67d49b4-fmw2f" event={"ID":"81360b3f-3235-4cc7-9ef3-e8f748037762","Type":"ContainerDied","Data":"e086a3c64a2da4fb8778731203db97137f5bdf9bf355be905b73134e8fd137d3"} Dec 05 11:06:55.021648 master-0 kubenswrapper[24928]: I1205 11:06:55.021554 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-54d5ff47b5-hzncp"] Dec 05 11:06:55.045636 master-0 kubenswrapper[24928]: I1205 11:06:55.045522 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-54d5ff47b5-hzncp"] Dec 05 11:06:55.454519 master-0 kubenswrapper[24928]: I1205 11:06:55.454262 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7c965f7db6-5tq2s"] Dec 05 11:06:56.017781 master-0 kubenswrapper[24928]: I1205 11:06:56.016774 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c965f7db6-5tq2s" event={"ID":"480d9df9-54a7-457d-8785-0c6074ea33f7","Type":"ContainerStarted","Data":"7dbe7ddd336f9effcd2d8d424c784fd4624d67ac5476e71a5e2da41d29167a14"} Dec 05 11:06:56.017781 master-0 kubenswrapper[24928]: I1205 11:06:56.016862 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c965f7db6-5tq2s" event={"ID":"480d9df9-54a7-457d-8785-0c6074ea33f7","Type":"ContainerStarted","Data":"401619032ba13494d372121c70f10ef33507ed10a81dfa5dc303a940d2bcf9ee"} Dec 05 11:06:56.220132 master-0 kubenswrapper[24928]: I1205 11:06:56.220072 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4540893-bc29-4b48-a741-ecb016eb0e73" path="/var/lib/kubelet/pods/f4540893-bc29-4b48-a741-ecb016eb0e73/volumes" Dec 05 11:06:57.032627 master-0 kubenswrapper[24928]: I1205 11:06:57.032567 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7c965f7db6-5tq2s" event={"ID":"480d9df9-54a7-457d-8785-0c6074ea33f7","Type":"ContainerStarted","Data":"a776c4843427e1d979d472157c854d4637805b83c86bdead0ebe85a5fdb2b442"} Dec 05 11:06:57.033665 master-0 kubenswrapper[24928]: I1205 11:06:57.032748 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:57.069865 master-0 kubenswrapper[24928]: I1205 11:06:57.069763 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7c965f7db6-5tq2s" podStartSLOduration=4.069744603 podStartE2EDuration="4.069744603s" podCreationTimestamp="2025-12-05 11:06:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:06:57.053899287 +0000 UTC m=+1177.057093138" watchObservedRunningTime="2025-12-05 11:06:57.069744603 +0000 UTC m=+1177.072938454" Dec 05 11:06:58.051155 master-0 kubenswrapper[24928]: I1205 11:06:58.051083 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:06:59.071105 master-0 kubenswrapper[24928]: I1205 11:06:59.071016 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-sppfx" event={"ID":"5183387d-f79b-4412-88ac-41b37866e07e","Type":"ContainerStarted","Data":"fd6274db687f6976b1c97618655aae11808c055a7cad0201a5eb5b0af9605762"} Dec 05 11:06:59.097622 master-0 kubenswrapper[24928]: I1205 11:06:59.097516 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-db-sync-sppfx" podStartSLOduration=4.124017266 podStartE2EDuration="7.097492319s" podCreationTimestamp="2025-12-05 11:06:52 +0000 UTC" firstStartedPulling="2025-12-05 11:06:54.159923733 +0000 UTC m=+1174.163117584" lastFinishedPulling="2025-12-05 11:06:57.133398786 +0000 UTC m=+1177.136592637" observedRunningTime="2025-12-05 11:06:59.091727778 +0000 UTC m=+1179.094921629" watchObservedRunningTime="2025-12-05 11:06:59.097492319 +0000 UTC m=+1179.100686170" Dec 05 11:07:01.113362 master-0 kubenswrapper[24928]: I1205 11:07:01.113295 24928 generic.go:334] "Generic (PLEG): container finished" podID="5183387d-f79b-4412-88ac-41b37866e07e" containerID="fd6274db687f6976b1c97618655aae11808c055a7cad0201a5eb5b0af9605762" exitCode=0 Dec 05 11:07:01.113938 master-0 kubenswrapper[24928]: I1205 11:07:01.113368 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-sppfx" event={"ID":"5183387d-f79b-4412-88ac-41b37866e07e","Type":"ContainerDied","Data":"fd6274db687f6976b1c97618655aae11808c055a7cad0201a5eb5b0af9605762"} Dec 05 11:07:04.777254 master-0 kubenswrapper[24928]: I1205 11:07:04.777178 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:07:04.778246 master-0 kubenswrapper[24928]: I1205 11:07:04.778197 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7c965f7db6-5tq2s" Dec 05 11:07:05.205342 master-0 kubenswrapper[24928]: I1205 11:07:05.205285 24928 scope.go:117] "RemoveContainer" containerID="1b0154660cc27dc1ba8da7a2dc97051ffd00f6f4aa5da15396421e670962c1d7" Dec 05 11:07:09.740054 master-0 kubenswrapper[24928]: I1205 11:07:09.740003 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:07:09.827032 master-0 kubenswrapper[24928]: I1205 11:07:09.826950 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"5183387d-f79b-4412-88ac-41b37866e07e\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " Dec 05 11:07:09.827032 master-0 kubenswrapper[24928]: I1205 11:07:09.827022 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wb5w8\" (UniqueName: \"kubernetes.io/projected/5183387d-f79b-4412-88ac-41b37866e07e-kube-api-access-wb5w8\") pod \"5183387d-f79b-4412-88ac-41b37866e07e\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " Dec 05 11:07:09.827287 master-0 kubenswrapper[24928]: I1205 11:07:09.827112 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic\") pod \"5183387d-f79b-4412-88ac-41b37866e07e\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " Dec 05 11:07:09.827287 master-0 kubenswrapper[24928]: I1205 11:07:09.827147 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-config\") pod \"5183387d-f79b-4412-88ac-41b37866e07e\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " Dec 05 11:07:09.827287 master-0 kubenswrapper[24928]: I1205 11:07:09.827209 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-scripts\") pod \"5183387d-f79b-4412-88ac-41b37866e07e\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " Dec 05 11:07:09.827287 master-0 kubenswrapper[24928]: I1205 11:07:09.827270 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5183387d-f79b-4412-88ac-41b37866e07e-etc-podinfo\") pod \"5183387d-f79b-4412-88ac-41b37866e07e\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " Dec 05 11:07:09.827413 master-0 kubenswrapper[24928]: I1205 11:07:09.827343 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-combined-ca-bundle\") pod \"5183387d-f79b-4412-88ac-41b37866e07e\" (UID: \"5183387d-f79b-4412-88ac-41b37866e07e\") " Dec 05 11:07:09.827539 master-0 kubenswrapper[24928]: I1205 11:07:09.827467 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic-inspector-dhcp-hostsdir" (OuterVolumeSpecName: "var-lib-ironic-inspector-dhcp-hostsdir") pod "5183387d-f79b-4412-88ac-41b37866e07e" (UID: "5183387d-f79b-4412-88ac-41b37866e07e"). InnerVolumeSpecName "var-lib-ironic-inspector-dhcp-hostsdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:07:09.827746 master-0 kubenswrapper[24928]: I1205 11:07:09.827686 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic" (OuterVolumeSpecName: "var-lib-ironic") pod "5183387d-f79b-4412-88ac-41b37866e07e" (UID: "5183387d-f79b-4412-88ac-41b37866e07e"). InnerVolumeSpecName "var-lib-ironic". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:07:09.828571 master-0 kubenswrapper[24928]: I1205 11:07:09.828532 24928 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic-inspector-dhcp-hostsdir\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:09.828571 master-0 kubenswrapper[24928]: I1205 11:07:09.828569 24928 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/5183387d-f79b-4412-88ac-41b37866e07e-var-lib-ironic\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:09.831333 master-0 kubenswrapper[24928]: I1205 11:07:09.831262 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-scripts" (OuterVolumeSpecName: "scripts") pod "5183387d-f79b-4412-88ac-41b37866e07e" (UID: "5183387d-f79b-4412-88ac-41b37866e07e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:09.831618 master-0 kubenswrapper[24928]: I1205 11:07:09.831570 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5183387d-f79b-4412-88ac-41b37866e07e-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "5183387d-f79b-4412-88ac-41b37866e07e" (UID: "5183387d-f79b-4412-88ac-41b37866e07e"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 11:07:09.832141 master-0 kubenswrapper[24928]: I1205 11:07:09.832092 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5183387d-f79b-4412-88ac-41b37866e07e-kube-api-access-wb5w8" (OuterVolumeSpecName: "kube-api-access-wb5w8") pod "5183387d-f79b-4412-88ac-41b37866e07e" (UID: "5183387d-f79b-4412-88ac-41b37866e07e"). InnerVolumeSpecName "kube-api-access-wb5w8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:09.863769 master-0 kubenswrapper[24928]: I1205 11:07:09.863300 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5183387d-f79b-4412-88ac-41b37866e07e" (UID: "5183387d-f79b-4412-88ac-41b37866e07e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:09.864383 master-0 kubenswrapper[24928]: I1205 11:07:09.864257 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-config" (OuterVolumeSpecName: "config") pod "5183387d-f79b-4412-88ac-41b37866e07e" (UID: "5183387d-f79b-4412-88ac-41b37866e07e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:09.930461 master-0 kubenswrapper[24928]: I1205 11:07:09.930285 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:09.930461 master-0 kubenswrapper[24928]: I1205 11:07:09.930340 24928 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5183387d-f79b-4412-88ac-41b37866e07e-etc-podinfo\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:09.930461 master-0 kubenswrapper[24928]: I1205 11:07:09.930355 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:09.930461 master-0 kubenswrapper[24928]: I1205 11:07:09.930367 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wb5w8\" (UniqueName: \"kubernetes.io/projected/5183387d-f79b-4412-88ac-41b37866e07e-kube-api-access-wb5w8\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:09.930461 master-0 kubenswrapper[24928]: I1205 11:07:09.930376 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5183387d-f79b-4412-88ac-41b37866e07e-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:10.220618 master-0 kubenswrapper[24928]: I1205 11:07:10.220007 24928 generic.go:334] "Generic (PLEG): container finished" podID="81360b3f-3235-4cc7-9ef3-e8f748037762" containerID="66554b0348b461b3687234b1ec9f9762acaa91069b3301f211e96317a7da9a8e" exitCode=0 Dec 05 11:07:10.228202 master-0 kubenswrapper[24928]: I1205 11:07:10.227403 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-sppfx" Dec 05 11:07:10.229128 master-0 kubenswrapper[24928]: I1205 11:07:10.229066 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd67d49b4-fmw2f" event={"ID":"81360b3f-3235-4cc7-9ef3-e8f748037762","Type":"ContainerDied","Data":"66554b0348b461b3687234b1ec9f9762acaa91069b3301f211e96317a7da9a8e"} Dec 05 11:07:10.229128 master-0 kubenswrapper[24928]: I1205 11:07:10.229122 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-sppfx" event={"ID":"5183387d-f79b-4412-88ac-41b37866e07e","Type":"ContainerDied","Data":"8a2ffae54ef8df93a24034a3fee4600ec48902cf8194e530070602b66a2f25aa"} Dec 05 11:07:10.229278 master-0 kubenswrapper[24928]: I1205 11:07:10.229140 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a2ffae54ef8df93a24034a3fee4600ec48902cf8194e530070602b66a2f25aa" Dec 05 11:07:11.791526 master-0 kubenswrapper[24928]: I1205 11:07:11.791449 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-g2gnb"] Dec 05 11:07:11.792101 master-0 kubenswrapper[24928]: E1205 11:07:11.792066 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5183387d-f79b-4412-88ac-41b37866e07e" containerName="ironic-inspector-db-sync" Dec 05 11:07:11.792101 master-0 kubenswrapper[24928]: I1205 11:07:11.792089 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="5183387d-f79b-4412-88ac-41b37866e07e" containerName="ironic-inspector-db-sync" Dec 05 11:07:11.792435 master-0 kubenswrapper[24928]: I1205 11:07:11.792395 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="5183387d-f79b-4412-88ac-41b37866e07e" containerName="ironic-inspector-db-sync" Dec 05 11:07:11.793820 master-0 kubenswrapper[24928]: I1205 11:07:11.793229 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g2gnb" Dec 05 11:07:11.825045 master-0 kubenswrapper[24928]: I1205 11:07:11.820397 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-g2gnb"] Dec 05 11:07:11.857907 master-0 kubenswrapper[24928]: I1205 11:07:11.854036 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-649rk"] Dec 05 11:07:11.857907 master-0 kubenswrapper[24928]: I1205 11:07:11.856598 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-649rk" Dec 05 11:07:11.898290 master-0 kubenswrapper[24928]: I1205 11:07:11.898205 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e73ed846-11a8-4c4b-b779-d7d362a3933f-operator-scripts\") pod \"nova-cell0-db-create-649rk\" (UID: \"e73ed846-11a8-4c4b-b779-d7d362a3933f\") " pod="openstack/nova-cell0-db-create-649rk" Dec 05 11:07:11.898290 master-0 kubenswrapper[24928]: I1205 11:07:11.898266 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06c47656-d5b2-405d-944d-3d4ce03be7b0-operator-scripts\") pod \"nova-api-db-create-g2gnb\" (UID: \"06c47656-d5b2-405d-944d-3d4ce03be7b0\") " pod="openstack/nova-api-db-create-g2gnb" Dec 05 11:07:11.898662 master-0 kubenswrapper[24928]: I1205 11:07:11.898373 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsggq\" (UniqueName: \"kubernetes.io/projected/06c47656-d5b2-405d-944d-3d4ce03be7b0-kube-api-access-fsggq\") pod \"nova-api-db-create-g2gnb\" (UID: \"06c47656-d5b2-405d-944d-3d4ce03be7b0\") " pod="openstack/nova-api-db-create-g2gnb" Dec 05 11:07:11.898927 master-0 kubenswrapper[24928]: I1205 11:07:11.898760 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlzkr\" (UniqueName: \"kubernetes.io/projected/e73ed846-11a8-4c4b-b779-d7d362a3933f-kube-api-access-dlzkr\") pod \"nova-cell0-db-create-649rk\" (UID: \"e73ed846-11a8-4c4b-b779-d7d362a3933f\") " pod="openstack/nova-cell0-db-create-649rk" Dec 05 11:07:11.947626 master-0 kubenswrapper[24928]: I1205 11:07:11.947524 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-649rk"] Dec 05 11:07:11.968493 master-0 kubenswrapper[24928]: I1205 11:07:11.967659 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-1c49-account-create-update-vwpvg"] Dec 05 11:07:11.970021 master-0 kubenswrapper[24928]: I1205 11:07:11.969919 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1c49-account-create-update-vwpvg" Dec 05 11:07:11.972403 master-0 kubenswrapper[24928]: I1205 11:07:11.972328 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 05 11:07:11.982974 master-0 kubenswrapper[24928]: I1205 11:07:11.982915 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1c49-account-create-update-vwpvg"] Dec 05 11:07:12.004561 master-0 kubenswrapper[24928]: I1205 11:07:12.004478 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpdfv\" (UniqueName: \"kubernetes.io/projected/f180288f-009b-4f64-bf07-543726ebc2e2-kube-api-access-tpdfv\") pod \"nova-api-1c49-account-create-update-vwpvg\" (UID: \"f180288f-009b-4f64-bf07-543726ebc2e2\") " pod="openstack/nova-api-1c49-account-create-update-vwpvg" Dec 05 11:07:12.004947 master-0 kubenswrapper[24928]: I1205 11:07:12.004630 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsggq\" (UniqueName: \"kubernetes.io/projected/06c47656-d5b2-405d-944d-3d4ce03be7b0-kube-api-access-fsggq\") pod \"nova-api-db-create-g2gnb\" (UID: \"06c47656-d5b2-405d-944d-3d4ce03be7b0\") " pod="openstack/nova-api-db-create-g2gnb" Dec 05 11:07:12.004947 master-0 kubenswrapper[24928]: I1205 11:07:12.004835 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlzkr\" (UniqueName: \"kubernetes.io/projected/e73ed846-11a8-4c4b-b779-d7d362a3933f-kube-api-access-dlzkr\") pod \"nova-cell0-db-create-649rk\" (UID: \"e73ed846-11a8-4c4b-b779-d7d362a3933f\") " pod="openstack/nova-cell0-db-create-649rk" Dec 05 11:07:12.005064 master-0 kubenswrapper[24928]: I1205 11:07:12.004971 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e73ed846-11a8-4c4b-b779-d7d362a3933f-operator-scripts\") pod \"nova-cell0-db-create-649rk\" (UID: \"e73ed846-11a8-4c4b-b779-d7d362a3933f\") " pod="openstack/nova-cell0-db-create-649rk" Dec 05 11:07:12.005064 master-0 kubenswrapper[24928]: I1205 11:07:12.004999 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06c47656-d5b2-405d-944d-3d4ce03be7b0-operator-scripts\") pod \"nova-api-db-create-g2gnb\" (UID: \"06c47656-d5b2-405d-944d-3d4ce03be7b0\") " pod="openstack/nova-api-db-create-g2gnb" Dec 05 11:07:12.005064 master-0 kubenswrapper[24928]: I1205 11:07:12.005026 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f180288f-009b-4f64-bf07-543726ebc2e2-operator-scripts\") pod \"nova-api-1c49-account-create-update-vwpvg\" (UID: \"f180288f-009b-4f64-bf07-543726ebc2e2\") " pod="openstack/nova-api-1c49-account-create-update-vwpvg" Dec 05 11:07:12.006173 master-0 kubenswrapper[24928]: I1205 11:07:12.006129 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06c47656-d5b2-405d-944d-3d4ce03be7b0-operator-scripts\") pod \"nova-api-db-create-g2gnb\" (UID: \"06c47656-d5b2-405d-944d-3d4ce03be7b0\") " pod="openstack/nova-api-db-create-g2gnb" Dec 05 11:07:12.010462 master-0 kubenswrapper[24928]: I1205 11:07:12.009125 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e73ed846-11a8-4c4b-b779-d7d362a3933f-operator-scripts\") pod \"nova-cell0-db-create-649rk\" (UID: \"e73ed846-11a8-4c4b-b779-d7d362a3933f\") " pod="openstack/nova-cell0-db-create-649rk" Dec 05 11:07:12.035476 master-0 kubenswrapper[24928]: I1205 11:07:12.024994 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlzkr\" (UniqueName: \"kubernetes.io/projected/e73ed846-11a8-4c4b-b779-d7d362a3933f-kube-api-access-dlzkr\") pod \"nova-cell0-db-create-649rk\" (UID: \"e73ed846-11a8-4c4b-b779-d7d362a3933f\") " pod="openstack/nova-cell0-db-create-649rk" Dec 05 11:07:12.035476 master-0 kubenswrapper[24928]: I1205 11:07:12.027713 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsggq\" (UniqueName: \"kubernetes.io/projected/06c47656-d5b2-405d-944d-3d4ce03be7b0-kube-api-access-fsggq\") pod \"nova-api-db-create-g2gnb\" (UID: \"06c47656-d5b2-405d-944d-3d4ce03be7b0\") " pod="openstack/nova-api-db-create-g2gnb" Dec 05 11:07:12.105718 master-0 kubenswrapper[24928]: I1205 11:07:12.104626 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-tqgcv"] Dec 05 11:07:12.116570 master-0 kubenswrapper[24928]: I1205 11:07:12.107824 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tqgcv" Dec 05 11:07:12.116570 master-0 kubenswrapper[24928]: I1205 11:07:12.111553 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f180288f-009b-4f64-bf07-543726ebc2e2-operator-scripts\") pod \"nova-api-1c49-account-create-update-vwpvg\" (UID: \"f180288f-009b-4f64-bf07-543726ebc2e2\") " pod="openstack/nova-api-1c49-account-create-update-vwpvg" Dec 05 11:07:12.116570 master-0 kubenswrapper[24928]: I1205 11:07:12.111787 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpdfv\" (UniqueName: \"kubernetes.io/projected/f180288f-009b-4f64-bf07-543726ebc2e2-kube-api-access-tpdfv\") pod \"nova-api-1c49-account-create-update-vwpvg\" (UID: \"f180288f-009b-4f64-bf07-543726ebc2e2\") " pod="openstack/nova-api-1c49-account-create-update-vwpvg" Dec 05 11:07:12.116570 master-0 kubenswrapper[24928]: I1205 11:07:12.115017 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f180288f-009b-4f64-bf07-543726ebc2e2-operator-scripts\") pod \"nova-api-1c49-account-create-update-vwpvg\" (UID: \"f180288f-009b-4f64-bf07-543726ebc2e2\") " pod="openstack/nova-api-1c49-account-create-update-vwpvg" Dec 05 11:07:12.132228 master-0 kubenswrapper[24928]: I1205 11:07:12.132159 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-2b0e-account-create-update-fpn9k"] Dec 05 11:07:12.133130 master-0 kubenswrapper[24928]: I1205 11:07:12.133068 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpdfv\" (UniqueName: \"kubernetes.io/projected/f180288f-009b-4f64-bf07-543726ebc2e2-kube-api-access-tpdfv\") pod \"nova-api-1c49-account-create-update-vwpvg\" (UID: \"f180288f-009b-4f64-bf07-543726ebc2e2\") " pod="openstack/nova-api-1c49-account-create-update-vwpvg" Dec 05 11:07:12.134527 master-0 kubenswrapper[24928]: I1205 11:07:12.134102 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" Dec 05 11:07:12.144625 master-0 kubenswrapper[24928]: I1205 11:07:12.144563 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 05 11:07:12.150712 master-0 kubenswrapper[24928]: I1205 11:07:12.146313 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g2gnb" Dec 05 11:07:12.158977 master-0 kubenswrapper[24928]: I1205 11:07:12.157658 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tqgcv"] Dec 05 11:07:12.184301 master-0 kubenswrapper[24928]: I1205 11:07:12.184229 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-649rk" Dec 05 11:07:12.198271 master-0 kubenswrapper[24928]: I1205 11:07:12.198211 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2b0e-account-create-update-fpn9k"] Dec 05 11:07:12.213570 master-0 kubenswrapper[24928]: I1205 11:07:12.213511 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mllp\" (UniqueName: \"kubernetes.io/projected/83a55bec-0267-4de3-97e8-8705a6056565-kube-api-access-4mllp\") pod \"nova-cell0-2b0e-account-create-update-fpn9k\" (UID: \"83a55bec-0267-4de3-97e8-8705a6056565\") " pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" Dec 05 11:07:12.213806 master-0 kubenswrapper[24928]: I1205 11:07:12.213601 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfvlv\" (UniqueName: \"kubernetes.io/projected/9c86a407-7dbd-49d1-87e8-c447bceb67bb-kube-api-access-jfvlv\") pod \"nova-cell1-db-create-tqgcv\" (UID: \"9c86a407-7dbd-49d1-87e8-c447bceb67bb\") " pod="openstack/nova-cell1-db-create-tqgcv" Dec 05 11:07:12.213806 master-0 kubenswrapper[24928]: I1205 11:07:12.213671 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c86a407-7dbd-49d1-87e8-c447bceb67bb-operator-scripts\") pod \"nova-cell1-db-create-tqgcv\" (UID: \"9c86a407-7dbd-49d1-87e8-c447bceb67bb\") " pod="openstack/nova-cell1-db-create-tqgcv" Dec 05 11:07:12.213806 master-0 kubenswrapper[24928]: I1205 11:07:12.213693 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83a55bec-0267-4de3-97e8-8705a6056565-operator-scripts\") pod \"nova-cell0-2b0e-account-create-update-fpn9k\" (UID: \"83a55bec-0267-4de3-97e8-8705a6056565\") " pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" Dec 05 11:07:12.279704 master-0 kubenswrapper[24928]: I1205 11:07:12.279609 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-9b34-account-create-update-gllvm"] Dec 05 11:07:12.281528 master-0 kubenswrapper[24928]: I1205 11:07:12.281487 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9b34-account-create-update-gllvm" Dec 05 11:07:12.284397 master-0 kubenswrapper[24928]: I1205 11:07:12.284207 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 05 11:07:12.290039 master-0 kubenswrapper[24928]: I1205 11:07:12.289971 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9b34-account-create-update-gllvm"] Dec 05 11:07:12.298201 master-0 kubenswrapper[24928]: I1205 11:07:12.298136 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1c49-account-create-update-vwpvg" Dec 05 11:07:12.321576 master-0 kubenswrapper[24928]: I1205 11:07:12.321519 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82b65c14-8968-437d-840d-30ccd28c57f9-operator-scripts\") pod \"nova-cell1-9b34-account-create-update-gllvm\" (UID: \"82b65c14-8968-437d-840d-30ccd28c57f9\") " pod="openstack/nova-cell1-9b34-account-create-update-gllvm" Dec 05 11:07:12.321817 master-0 kubenswrapper[24928]: I1205 11:07:12.321721 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mllp\" (UniqueName: \"kubernetes.io/projected/83a55bec-0267-4de3-97e8-8705a6056565-kube-api-access-4mllp\") pod \"nova-cell0-2b0e-account-create-update-fpn9k\" (UID: \"83a55bec-0267-4de3-97e8-8705a6056565\") " pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" Dec 05 11:07:12.322590 master-0 kubenswrapper[24928]: I1205 11:07:12.322147 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2vtl\" (UniqueName: \"kubernetes.io/projected/82b65c14-8968-437d-840d-30ccd28c57f9-kube-api-access-z2vtl\") pod \"nova-cell1-9b34-account-create-update-gllvm\" (UID: \"82b65c14-8968-437d-840d-30ccd28c57f9\") " pod="openstack/nova-cell1-9b34-account-create-update-gllvm" Dec 05 11:07:12.322590 master-0 kubenswrapper[24928]: I1205 11:07:12.322239 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jfvlv\" (UniqueName: \"kubernetes.io/projected/9c86a407-7dbd-49d1-87e8-c447bceb67bb-kube-api-access-jfvlv\") pod \"nova-cell1-db-create-tqgcv\" (UID: \"9c86a407-7dbd-49d1-87e8-c447bceb67bb\") " pod="openstack/nova-cell1-db-create-tqgcv" Dec 05 11:07:12.323433 master-0 kubenswrapper[24928]: I1205 11:07:12.323377 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c86a407-7dbd-49d1-87e8-c447bceb67bb-operator-scripts\") pod \"nova-cell1-db-create-tqgcv\" (UID: \"9c86a407-7dbd-49d1-87e8-c447bceb67bb\") " pod="openstack/nova-cell1-db-create-tqgcv" Dec 05 11:07:12.323675 master-0 kubenswrapper[24928]: I1205 11:07:12.323633 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83a55bec-0267-4de3-97e8-8705a6056565-operator-scripts\") pod \"nova-cell0-2b0e-account-create-update-fpn9k\" (UID: \"83a55bec-0267-4de3-97e8-8705a6056565\") " pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" Dec 05 11:07:12.323745 master-0 kubenswrapper[24928]: I1205 11:07:12.323669 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c86a407-7dbd-49d1-87e8-c447bceb67bb-operator-scripts\") pod \"nova-cell1-db-create-tqgcv\" (UID: \"9c86a407-7dbd-49d1-87e8-c447bceb67bb\") " pod="openstack/nova-cell1-db-create-tqgcv" Dec 05 11:07:12.324800 master-0 kubenswrapper[24928]: I1205 11:07:12.324750 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83a55bec-0267-4de3-97e8-8705a6056565-operator-scripts\") pod \"nova-cell0-2b0e-account-create-update-fpn9k\" (UID: \"83a55bec-0267-4de3-97e8-8705a6056565\") " pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" Dec 05 11:07:12.342054 master-0 kubenswrapper[24928]: I1205 11:07:12.341991 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mllp\" (UniqueName: \"kubernetes.io/projected/83a55bec-0267-4de3-97e8-8705a6056565-kube-api-access-4mllp\") pod \"nova-cell0-2b0e-account-create-update-fpn9k\" (UID: \"83a55bec-0267-4de3-97e8-8705a6056565\") " pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" Dec 05 11:07:12.344196 master-0 kubenswrapper[24928]: I1205 11:07:12.344131 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfvlv\" (UniqueName: \"kubernetes.io/projected/9c86a407-7dbd-49d1-87e8-c447bceb67bb-kube-api-access-jfvlv\") pod \"nova-cell1-db-create-tqgcv\" (UID: \"9c86a407-7dbd-49d1-87e8-c447bceb67bb\") " pod="openstack/nova-cell1-db-create-tqgcv" Dec 05 11:07:12.426977 master-0 kubenswrapper[24928]: I1205 11:07:12.426777 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2vtl\" (UniqueName: \"kubernetes.io/projected/82b65c14-8968-437d-840d-30ccd28c57f9-kube-api-access-z2vtl\") pod \"nova-cell1-9b34-account-create-update-gllvm\" (UID: \"82b65c14-8968-437d-840d-30ccd28c57f9\") " pod="openstack/nova-cell1-9b34-account-create-update-gllvm" Dec 05 11:07:12.427196 master-0 kubenswrapper[24928]: I1205 11:07:12.427057 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82b65c14-8968-437d-840d-30ccd28c57f9-operator-scripts\") pod \"nova-cell1-9b34-account-create-update-gllvm\" (UID: \"82b65c14-8968-437d-840d-30ccd28c57f9\") " pod="openstack/nova-cell1-9b34-account-create-update-gllvm" Dec 05 11:07:12.428035 master-0 kubenswrapper[24928]: I1205 11:07:12.427990 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82b65c14-8968-437d-840d-30ccd28c57f9-operator-scripts\") pod \"nova-cell1-9b34-account-create-update-gllvm\" (UID: \"82b65c14-8968-437d-840d-30ccd28c57f9\") " pod="openstack/nova-cell1-9b34-account-create-update-gllvm" Dec 05 11:07:12.447639 master-0 kubenswrapper[24928]: I1205 11:07:12.447592 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2vtl\" (UniqueName: \"kubernetes.io/projected/82b65c14-8968-437d-840d-30ccd28c57f9-kube-api-access-z2vtl\") pod \"nova-cell1-9b34-account-create-update-gllvm\" (UID: \"82b65c14-8968-437d-840d-30ccd28c57f9\") " pod="openstack/nova-cell1-9b34-account-create-update-gllvm" Dec 05 11:07:12.555660 master-0 kubenswrapper[24928]: I1205 11:07:12.554784 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tqgcv" Dec 05 11:07:12.563945 master-0 kubenswrapper[24928]: I1205 11:07:12.562644 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" Dec 05 11:07:12.610147 master-0 kubenswrapper[24928]: I1205 11:07:12.610063 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9b34-account-create-update-gllvm" Dec 05 11:07:13.748754 master-0 kubenswrapper[24928]: I1205 11:07:13.748680 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:07:13.749331 master-0 kubenswrapper[24928]: I1205 11:07:13.749068 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-ec941-default-internal-api-0" podUID="fddf0e6c-7977-4429-9015-732fd8dbb6a1" containerName="glance-log" containerID="cri-o://7c116d246b2d3f460a0aa28456f2c1fe5c53a3c50f6ed7a5a3f758d1ad2cb7c5" gracePeriod=30 Dec 05 11:07:13.749865 master-0 kubenswrapper[24928]: I1205 11:07:13.749817 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-ec941-default-internal-api-0" podUID="fddf0e6c-7977-4429-9015-732fd8dbb6a1" containerName="glance-httpd" containerID="cri-o://2130b735befea54638f55b346f56e9884c2c0ab93e77888bb7dad2d68c56bea0" gracePeriod=30 Dec 05 11:07:14.140477 master-0 kubenswrapper[24928]: I1205 11:07:14.135787 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68cb7b79bf-rs4hj"] Dec 05 11:07:14.218334 master-0 kubenswrapper[24928]: I1205 11:07:14.203346 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:07:14.218334 master-0 kubenswrapper[24928]: I1205 11:07:14.210722 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.366503 master-0 kubenswrapper[24928]: I1205 11:07:14.363855 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-httpd-config\") pod \"81360b3f-3235-4cc7-9ef3-e8f748037762\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " Dec 05 11:07:14.366503 master-0 kubenswrapper[24928]: I1205 11:07:14.364012 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfkp5\" (UniqueName: \"kubernetes.io/projected/81360b3f-3235-4cc7-9ef3-e8f748037762-kube-api-access-nfkp5\") pod \"81360b3f-3235-4cc7-9ef3-e8f748037762\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " Dec 05 11:07:14.366503 master-0 kubenswrapper[24928]: I1205 11:07:14.364109 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-combined-ca-bundle\") pod \"81360b3f-3235-4cc7-9ef3-e8f748037762\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " Dec 05 11:07:14.366503 master-0 kubenswrapper[24928]: I1205 11:07:14.364229 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-config\") pod \"81360b3f-3235-4cc7-9ef3-e8f748037762\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " Dec 05 11:07:14.366503 master-0 kubenswrapper[24928]: I1205 11:07:14.364282 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-ovndb-tls-certs\") pod \"81360b3f-3235-4cc7-9ef3-e8f748037762\" (UID: \"81360b3f-3235-4cc7-9ef3-e8f748037762\") " Dec 05 11:07:14.366503 master-0 kubenswrapper[24928]: I1205 11:07:14.364664 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-nb\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.366503 master-0 kubenswrapper[24928]: I1205 11:07:14.364715 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-config\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.366503 master-0 kubenswrapper[24928]: I1205 11:07:14.364795 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmbcg\" (UniqueName: \"kubernetes.io/projected/b7a38786-cb62-46b5-8a00-54f6d01545b8-kube-api-access-hmbcg\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.366503 master-0 kubenswrapper[24928]: I1205 11:07:14.365026 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-sb\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.366503 master-0 kubenswrapper[24928]: I1205 11:07:14.365072 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-svc\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.366503 master-0 kubenswrapper[24928]: I1205 11:07:14.365137 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-swift-storage-0\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.370092 master-0 kubenswrapper[24928]: I1205 11:07:14.369781 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "81360b3f-3235-4cc7-9ef3-e8f748037762" (UID: "81360b3f-3235-4cc7-9ef3-e8f748037762"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:14.409552 master-0 kubenswrapper[24928]: I1205 11:07:14.409482 24928 generic.go:334] "Generic (PLEG): container finished" podID="fddf0e6c-7977-4429-9015-732fd8dbb6a1" containerID="7c116d246b2d3f460a0aa28456f2c1fe5c53a3c50f6ed7a5a3f758d1ad2cb7c5" exitCode=143 Dec 05 11:07:14.409890 master-0 kubenswrapper[24928]: I1205 11:07:14.409817 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81360b3f-3235-4cc7-9ef3-e8f748037762-kube-api-access-nfkp5" (OuterVolumeSpecName: "kube-api-access-nfkp5") pod "81360b3f-3235-4cc7-9ef3-e8f748037762" (UID: "81360b3f-3235-4cc7-9ef3-e8f748037762"). InnerVolumeSpecName "kube-api-access-nfkp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:14.412822 master-0 kubenswrapper[24928]: I1205 11:07:14.412762 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7bd67d49b4-fmw2f" Dec 05 11:07:14.475461 master-0 kubenswrapper[24928]: I1205 11:07:14.471675 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-svc\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.475461 master-0 kubenswrapper[24928]: I1205 11:07:14.471776 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-swift-storage-0\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.475461 master-0 kubenswrapper[24928]: I1205 11:07:14.471928 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-nb\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.475461 master-0 kubenswrapper[24928]: I1205 11:07:14.471962 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-config\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.475461 master-0 kubenswrapper[24928]: I1205 11:07:14.472062 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmbcg\" (UniqueName: \"kubernetes.io/projected/b7a38786-cb62-46b5-8a00-54f6d01545b8-kube-api-access-hmbcg\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.475461 master-0 kubenswrapper[24928]: I1205 11:07:14.472169 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-sb\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.475461 master-0 kubenswrapper[24928]: I1205 11:07:14.472231 24928 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-httpd-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:14.475461 master-0 kubenswrapper[24928]: I1205 11:07:14.472249 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfkp5\" (UniqueName: \"kubernetes.io/projected/81360b3f-3235-4cc7-9ef3-e8f748037762-kube-api-access-nfkp5\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:14.475461 master-0 kubenswrapper[24928]: I1205 11:07:14.474648 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-sb\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.475461 master-0 kubenswrapper[24928]: I1205 11:07:14.474829 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-nb\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.476291 master-0 kubenswrapper[24928]: I1205 11:07:14.475818 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-config\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.478531 master-0 kubenswrapper[24928]: I1205 11:07:14.477967 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-swift-storage-0\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.478531 master-0 kubenswrapper[24928]: I1205 11:07:14.475490 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-svc\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.510783 master-0 kubenswrapper[24928]: I1205 11:07:14.510232 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmbcg\" (UniqueName: \"kubernetes.io/projected/b7a38786-cb62-46b5-8a00-54f6d01545b8-kube-api-access-hmbcg\") pod \"dnsmasq-dns-68cb7b79bf-rs4hj\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.538828 master-0 kubenswrapper[24928]: I1205 11:07:14.538730 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-config" (OuterVolumeSpecName: "config") pod "81360b3f-3235-4cc7-9ef3-e8f748037762" (UID: "81360b3f-3235-4cc7-9ef3-e8f748037762"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:14.574223 master-0 kubenswrapper[24928]: I1205 11:07:14.574178 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:14.575274 master-0 kubenswrapper[24928]: I1205 11:07:14.575201 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:14.621444 master-0 kubenswrapper[24928]: I1205 11:07:14.617911 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "81360b3f-3235-4cc7-9ef3-e8f748037762" (UID: "81360b3f-3235-4cc7-9ef3-e8f748037762"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:14.676734 master-0 kubenswrapper[24928]: I1205 11:07:14.676631 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "81360b3f-3235-4cc7-9ef3-e8f748037762" (UID: "81360b3f-3235-4cc7-9ef3-e8f748037762"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:14.690751 master-0 kubenswrapper[24928]: I1205 11:07:14.689176 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:14.690751 master-0 kubenswrapper[24928]: I1205 11:07:14.689242 24928 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81360b3f-3235-4cc7-9ef3-e8f748037762-ovndb-tls-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:15.012872 master-0 kubenswrapper[24928]: I1205 11:07:15.012804 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"fddf0e6c-7977-4429-9015-732fd8dbb6a1","Type":"ContainerDied","Data":"7c116d246b2d3f460a0aa28456f2c1fe5c53a3c50f6ed7a5a3f758d1ad2cb7c5"} Dec 05 11:07:15.012872 master-0 kubenswrapper[24928]: I1205 11:07:15.012864 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7bd67d49b4-fmw2f" event={"ID":"81360b3f-3235-4cc7-9ef3-e8f748037762","Type":"ContainerDied","Data":"9db0491738dd6eeec3d1bb79bd094595b7517436950f2ab2ff8e9181b13fba21"} Dec 05 11:07:15.013552 master-0 kubenswrapper[24928]: I1205 11:07:15.012946 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68cb7b79bf-rs4hj"] Dec 05 11:07:15.013552 master-0 kubenswrapper[24928]: I1205 11:07:15.012963 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-0"] Dec 05 11:07:15.013552 master-0 kubenswrapper[24928]: E1205 11:07:15.013400 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81360b3f-3235-4cc7-9ef3-e8f748037762" containerName="neutron-api" Dec 05 11:07:15.013552 master-0 kubenswrapper[24928]: I1205 11:07:15.013413 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="81360b3f-3235-4cc7-9ef3-e8f748037762" containerName="neutron-api" Dec 05 11:07:15.013552 master-0 kubenswrapper[24928]: E1205 11:07:15.013491 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81360b3f-3235-4cc7-9ef3-e8f748037762" containerName="neutron-httpd" Dec 05 11:07:15.013552 master-0 kubenswrapper[24928]: I1205 11:07:15.013498 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="81360b3f-3235-4cc7-9ef3-e8f748037762" containerName="neutron-httpd" Dec 05 11:07:15.013853 master-0 kubenswrapper[24928]: I1205 11:07:15.013830 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="81360b3f-3235-4cc7-9ef3-e8f748037762" containerName="neutron-httpd" Dec 05 11:07:15.013853 master-0 kubenswrapper[24928]: I1205 11:07:15.013854 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="81360b3f-3235-4cc7-9ef3-e8f748037762" containerName="neutron-api" Dec 05 11:07:15.020590 master-0 kubenswrapper[24928]: I1205 11:07:15.017205 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Dec 05 11:07:15.020590 master-0 kubenswrapper[24928]: I1205 11:07:15.017315 24928 scope.go:117] "RemoveContainer" containerID="e086a3c64a2da4fb8778731203db97137f5bdf9bf355be905b73134e8fd137d3" Dec 05 11:07:15.020590 master-0 kubenswrapper[24928]: I1205 11:07:15.017339 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Dec 05 11:07:15.033821 master-0 kubenswrapper[24928]: I1205 11:07:15.023608 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-transport-url-ironic-inspector-transport" Dec 05 11:07:15.033821 master-0 kubenswrapper[24928]: I1205 11:07:15.024169 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Dec 05 11:07:15.033821 master-0 kubenswrapper[24928]: I1205 11:07:15.024358 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Dec 05 11:07:15.099878 master-0 kubenswrapper[24928]: I1205 11:07:15.099807 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-scripts\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.099878 master-0 kubenswrapper[24928]: I1205 11:07:15.099877 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.100057 master-0 kubenswrapper[24928]: I1205 11:07:15.099902 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-config\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.100316 master-0 kubenswrapper[24928]: I1205 11:07:15.100231 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm5lr\" (UniqueName: \"kubernetes.io/projected/313a9b9c-4454-4927-a218-761e791cf156-kube-api-access-pm5lr\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.100629 master-0 kubenswrapper[24928]: I1205 11:07:15.100579 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.100816 master-0 kubenswrapper[24928]: I1205 11:07:15.100759 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.101322 master-0 kubenswrapper[24928]: I1205 11:07:15.101286 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/313a9b9c-4454-4927-a218-761e791cf156-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.103691 master-0 kubenswrapper[24928]: I1205 11:07:15.101960 24928 scope.go:117] "RemoveContainer" containerID="66554b0348b461b3687234b1ec9f9762acaa91069b3301f211e96317a7da9a8e" Dec 05 11:07:15.213635 master-0 kubenswrapper[24928]: I1205 11:07:15.213549 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm5lr\" (UniqueName: \"kubernetes.io/projected/313a9b9c-4454-4927-a218-761e791cf156-kube-api-access-pm5lr\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.213830 master-0 kubenswrapper[24928]: I1205 11:07:15.213666 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.213830 master-0 kubenswrapper[24928]: I1205 11:07:15.213739 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.213983 master-0 kubenswrapper[24928]: I1205 11:07:15.213949 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/313a9b9c-4454-4927-a218-761e791cf156-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.214166 master-0 kubenswrapper[24928]: I1205 11:07:15.214140 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-scripts\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.214362 master-0 kubenswrapper[24928]: I1205 11:07:15.214177 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.214515 master-0 kubenswrapper[24928]: I1205 11:07:15.214464 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.215958 master-0 kubenswrapper[24928]: I1205 11:07:15.215927 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-config\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.216307 master-0 kubenswrapper[24928]: I1205 11:07:15.216250 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.220568 master-0 kubenswrapper[24928]: I1205 11:07:15.220517 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-scripts\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.221220 master-0 kubenswrapper[24928]: I1205 11:07:15.221179 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-config\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.221382 master-0 kubenswrapper[24928]: I1205 11:07:15.221337 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/313a9b9c-4454-4927-a218-761e791cf156-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.237859 master-0 kubenswrapper[24928]: I1205 11:07:15.237312 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.243005 master-0 kubenswrapper[24928]: I1205 11:07:15.242922 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm5lr\" (UniqueName: \"kubernetes.io/projected/313a9b9c-4454-4927-a218-761e791cf156-kube-api-access-pm5lr\") pod \"ironic-inspector-0\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:15.244754 master-0 kubenswrapper[24928]: I1205 11:07:15.244699 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7bd67d49b4-fmw2f"] Dec 05 11:07:15.259266 master-0 kubenswrapper[24928]: I1205 11:07:15.259182 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7bd67d49b4-fmw2f"] Dec 05 11:07:15.274697 master-0 kubenswrapper[24928]: I1205 11:07:15.274404 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9b34-account-create-update-gllvm"] Dec 05 11:07:15.460564 master-0 kubenswrapper[24928]: I1205 11:07:15.459768 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"fa285890-6077-4460-8935-e683d1c36783","Type":"ContainerStarted","Data":"c8ae43e4df17cfb8f40d30a9ebe6c764df5b03c0ba09e1546117cb9e6adf5a4a"} Dec 05 11:07:15.470634 master-0 kubenswrapper[24928]: I1205 11:07:15.464725 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" event={"ID":"2a9f99ee-4977-4963-a6b1-2a2bd2e849d1","Type":"ContainerStarted","Data":"2f1bc437c5be7c8fc398c8798b34d7d2b4791e637da6367cbcb3a82753a298aa"} Dec 05 11:07:15.470634 master-0 kubenswrapper[24928]: I1205 11:07:15.464957 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:07:15.478988 master-0 kubenswrapper[24928]: I1205 11:07:15.478154 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9b34-account-create-update-gllvm" event={"ID":"82b65c14-8968-437d-840d-30ccd28c57f9","Type":"ContainerStarted","Data":"da5a0fa9aa6942984dd3451c39990e1cb3614bdaabb8f285637ac1a90ccbf20b"} Dec 05 11:07:15.491040 master-0 kubenswrapper[24928]: I1205 11:07:15.490945 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Dec 05 11:07:15.491671 master-0 kubenswrapper[24928]: I1205 11:07:15.491577 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"2d8079b6-0de1-4ffd-b008-9d74bf9673ae","Type":"ContainerStarted","Data":"372563fe24f0f6f0192af404a2a04904a2370849e44934fb43245471fcdb49ff"} Dec 05 11:07:15.500701 master-0 kubenswrapper[24928]: I1205 11:07:15.499408 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=5.062074217 podStartE2EDuration="25.499382824s" podCreationTimestamp="2025-12-05 11:06:50 +0000 UTC" firstStartedPulling="2025-12-05 11:06:53.911109475 +0000 UTC m=+1173.914303326" lastFinishedPulling="2025-12-05 11:07:14.348418082 +0000 UTC m=+1194.351611933" observedRunningTime="2025-12-05 11:07:15.490788384 +0000 UTC m=+1195.493982235" watchObservedRunningTime="2025-12-05 11:07:15.499382824 +0000 UTC m=+1195.502576675" Dec 05 11:07:15.715247 master-0 kubenswrapper[24928]: I1205 11:07:15.715158 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:07:15.715712 master-0 kubenswrapper[24928]: I1205 11:07:15.715635 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-ec941-default-external-api-0" podUID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerName="glance-log" containerID="cri-o://73bb8e24a58ffbe2967e4ba93ebbac3bff4baac27f7243d2c461aaa0802ea5d2" gracePeriod=30 Dec 05 11:07:15.716277 master-0 kubenswrapper[24928]: I1205 11:07:15.716200 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-ec941-default-external-api-0" podUID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerName="glance-httpd" containerID="cri-o://acd53d82e3153e795b5309b7c86275854aea6e0fe04723a35820bec9073a5ad8" gracePeriod=30 Dec 05 11:07:15.762244 master-0 kubenswrapper[24928]: I1205 11:07:15.752294 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1c49-account-create-update-vwpvg"] Dec 05 11:07:15.764806 master-0 kubenswrapper[24928]: I1205 11:07:15.764087 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-649rk"] Dec 05 11:07:15.775529 master-0 kubenswrapper[24928]: I1205 11:07:15.775314 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2b0e-account-create-update-fpn9k"] Dec 05 11:07:15.953870 master-0 kubenswrapper[24928]: I1205 11:07:15.953761 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-g2gnb"] Dec 05 11:07:15.969784 master-0 kubenswrapper[24928]: I1205 11:07:15.968344 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tqgcv"] Dec 05 11:07:15.979755 master-0 kubenswrapper[24928]: I1205 11:07:15.979685 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68cb7b79bf-rs4hj"] Dec 05 11:07:16.257777 master-0 kubenswrapper[24928]: I1205 11:07:16.256211 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81360b3f-3235-4cc7-9ef3-e8f748037762" path="/var/lib/kubelet/pods/81360b3f-3235-4cc7-9ef3-e8f748037762/volumes" Dec 05 11:07:16.271758 master-0 kubenswrapper[24928]: I1205 11:07:16.271697 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Dec 05 11:07:16.510149 master-0 kubenswrapper[24928]: I1205 11:07:16.509624 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"313a9b9c-4454-4927-a218-761e791cf156","Type":"ContainerStarted","Data":"f12070fccf429dd91d728ac7daffd4a8fbe51645b1b9b50742a02ae76562f9ba"} Dec 05 11:07:16.512966 master-0 kubenswrapper[24928]: I1205 11:07:16.512537 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" event={"ID":"83a55bec-0267-4de3-97e8-8705a6056565","Type":"ContainerStarted","Data":"9fb6867aa7cfa7925e04c045e72c2da1fe1c373a0b7241f16eb9a30adbcfcf11"} Dec 05 11:07:16.513056 master-0 kubenswrapper[24928]: I1205 11:07:16.512955 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" event={"ID":"83a55bec-0267-4de3-97e8-8705a6056565","Type":"ContainerStarted","Data":"241e80b1b0c15d2da03ed575e6017698f627610a44995dbdc95e2691ff98624e"} Dec 05 11:07:16.515159 master-0 kubenswrapper[24928]: I1205 11:07:16.515096 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tqgcv" event={"ID":"9c86a407-7dbd-49d1-87e8-c447bceb67bb","Type":"ContainerStarted","Data":"0b891267e1e08f06ff8f2233654172342f8078c9cf3a0e2fc711005a5b723ba8"} Dec 05 11:07:16.515159 master-0 kubenswrapper[24928]: I1205 11:07:16.515153 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tqgcv" event={"ID":"9c86a407-7dbd-49d1-87e8-c447bceb67bb","Type":"ContainerStarted","Data":"b706a0c33fc18345c5b0f3f1d8ed5d311a2781bad2dc17e66152d9f3811a8bc0"} Dec 05 11:07:16.521580 master-0 kubenswrapper[24928]: I1205 11:07:16.521522 24928 generic.go:334] "Generic (PLEG): container finished" podID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerID="73bb8e24a58ffbe2967e4ba93ebbac3bff4baac27f7243d2c461aaa0802ea5d2" exitCode=143 Dec 05 11:07:16.521687 master-0 kubenswrapper[24928]: I1205 11:07:16.521652 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"e1a0529b-71c2-4a72-b75f-0b56b07e7c64","Type":"ContainerDied","Data":"73bb8e24a58ffbe2967e4ba93ebbac3bff4baac27f7243d2c461aaa0802ea5d2"} Dec 05 11:07:16.524691 master-0 kubenswrapper[24928]: I1205 11:07:16.524606 24928 generic.go:334] "Generic (PLEG): container finished" podID="e73ed846-11a8-4c4b-b779-d7d362a3933f" containerID="acce575d949e3045ed4bd69b5f14b9260c305ebba6ba3756fcaf6c5db9f0ad61" exitCode=0 Dec 05 11:07:16.524837 master-0 kubenswrapper[24928]: I1205 11:07:16.524800 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-649rk" event={"ID":"e73ed846-11a8-4c4b-b779-d7d362a3933f","Type":"ContainerDied","Data":"acce575d949e3045ed4bd69b5f14b9260c305ebba6ba3756fcaf6c5db9f0ad61"} Dec 05 11:07:16.524880 master-0 kubenswrapper[24928]: I1205 11:07:16.524837 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-649rk" event={"ID":"e73ed846-11a8-4c4b-b779-d7d362a3933f","Type":"ContainerStarted","Data":"a07d4f0dbb9d6e35e5c60d8f36d62c53539e2e1faf1b99fb77c1f9eac54c5527"} Dec 05 11:07:16.526691 master-0 kubenswrapper[24928]: I1205 11:07:16.526645 24928 generic.go:334] "Generic (PLEG): container finished" podID="f180288f-009b-4f64-bf07-543726ebc2e2" containerID="aa96d63f095df2ca19ce1477b28f29c86bdcdf74b8f2261ad3dcd4bc3f93b052" exitCode=0 Dec 05 11:07:16.526757 master-0 kubenswrapper[24928]: I1205 11:07:16.526719 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1c49-account-create-update-vwpvg" event={"ID":"f180288f-009b-4f64-bf07-543726ebc2e2","Type":"ContainerDied","Data":"aa96d63f095df2ca19ce1477b28f29c86bdcdf74b8f2261ad3dcd4bc3f93b052"} Dec 05 11:07:16.526757 master-0 kubenswrapper[24928]: I1205 11:07:16.526747 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1c49-account-create-update-vwpvg" event={"ID":"f180288f-009b-4f64-bf07-543726ebc2e2","Type":"ContainerStarted","Data":"c609a74895d38f9693fc867343f0c4d6fac9f3cb2a4e895602733af49f1981ec"} Dec 05 11:07:16.528444 master-0 kubenswrapper[24928]: I1205 11:07:16.528378 24928 generic.go:334] "Generic (PLEG): container finished" podID="82b65c14-8968-437d-840d-30ccd28c57f9" containerID="f459c0129b9cd0a94447984332462d3ad67721492a24a34df72b8014cf294a34" exitCode=0 Dec 05 11:07:16.528526 master-0 kubenswrapper[24928]: I1205 11:07:16.528463 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9b34-account-create-update-gllvm" event={"ID":"82b65c14-8968-437d-840d-30ccd28c57f9","Type":"ContainerDied","Data":"f459c0129b9cd0a94447984332462d3ad67721492a24a34df72b8014cf294a34"} Dec 05 11:07:16.534285 master-0 kubenswrapper[24928]: I1205 11:07:16.534198 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" podStartSLOduration=4.534179042 podStartE2EDuration="4.534179042s" podCreationTimestamp="2025-12-05 11:07:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:07:16.532764357 +0000 UTC m=+1196.535958208" watchObservedRunningTime="2025-12-05 11:07:16.534179042 +0000 UTC m=+1196.537372893" Dec 05 11:07:16.536034 master-0 kubenswrapper[24928]: I1205 11:07:16.535997 24928 generic.go:334] "Generic (PLEG): container finished" podID="b7a38786-cb62-46b5-8a00-54f6d01545b8" containerID="ebfbe2252b2a4ef8d0f1b9c6c4eb2b0091893ac0af60288bcc8b25ac01c6913c" exitCode=0 Dec 05 11:07:16.536117 master-0 kubenswrapper[24928]: I1205 11:07:16.536070 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" event={"ID":"b7a38786-cb62-46b5-8a00-54f6d01545b8","Type":"ContainerDied","Data":"ebfbe2252b2a4ef8d0f1b9c6c4eb2b0091893ac0af60288bcc8b25ac01c6913c"} Dec 05 11:07:16.536161 master-0 kubenswrapper[24928]: I1205 11:07:16.536126 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" event={"ID":"b7a38786-cb62-46b5-8a00-54f6d01545b8","Type":"ContainerStarted","Data":"d49450deb614190049128fc15b51a69553abc017a36e43274fc10ae71124834a"} Dec 05 11:07:16.546995 master-0 kubenswrapper[24928]: I1205 11:07:16.546913 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g2gnb" event={"ID":"06c47656-d5b2-405d-944d-3d4ce03be7b0","Type":"ContainerStarted","Data":"4e24ad44d524026003f4888853d5d3caeac4e3f2ac890101297980e102d96974"} Dec 05 11:07:16.547109 master-0 kubenswrapper[24928]: I1205 11:07:16.546995 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g2gnb" event={"ID":"06c47656-d5b2-405d-944d-3d4ce03be7b0","Type":"ContainerStarted","Data":"38cf6d9229224b5097ccca695b90e356df9d189e14979f2fca3757c20f7af0b2"} Dec 05 11:07:16.565378 master-0 kubenswrapper[24928]: I1205 11:07:16.565220 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-tqgcv" podStartSLOduration=4.565190468 podStartE2EDuration="4.565190468s" podCreationTimestamp="2025-12-05 11:07:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:07:16.55748015 +0000 UTC m=+1196.560674001" watchObservedRunningTime="2025-12-05 11:07:16.565190468 +0000 UTC m=+1196.568384319" Dec 05 11:07:18.157395 master-0 kubenswrapper[24928]: I1205 11:07:18.157302 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1c49-account-create-update-vwpvg" Dec 05 11:07:18.284011 master-0 kubenswrapper[24928]: I1205 11:07:18.283955 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-649rk" Dec 05 11:07:18.290226 master-0 kubenswrapper[24928]: I1205 11:07:18.290186 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9b34-account-create-update-gllvm" Dec 05 11:07:18.323187 master-0 kubenswrapper[24928]: I1205 11:07:18.323102 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpdfv\" (UniqueName: \"kubernetes.io/projected/f180288f-009b-4f64-bf07-543726ebc2e2-kube-api-access-tpdfv\") pod \"f180288f-009b-4f64-bf07-543726ebc2e2\" (UID: \"f180288f-009b-4f64-bf07-543726ebc2e2\") " Dec 05 11:07:18.323416 master-0 kubenswrapper[24928]: I1205 11:07:18.323215 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f180288f-009b-4f64-bf07-543726ebc2e2-operator-scripts\") pod \"f180288f-009b-4f64-bf07-543726ebc2e2\" (UID: \"f180288f-009b-4f64-bf07-543726ebc2e2\") " Dec 05 11:07:18.324019 master-0 kubenswrapper[24928]: I1205 11:07:18.323982 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f180288f-009b-4f64-bf07-543726ebc2e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f180288f-009b-4f64-bf07-543726ebc2e2" (UID: "f180288f-009b-4f64-bf07-543726ebc2e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:07:18.326085 master-0 kubenswrapper[24928]: I1205 11:07:18.326034 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f180288f-009b-4f64-bf07-543726ebc2e2-kube-api-access-tpdfv" (OuterVolumeSpecName: "kube-api-access-tpdfv") pod "f180288f-009b-4f64-bf07-543726ebc2e2" (UID: "f180288f-009b-4f64-bf07-543726ebc2e2"). InnerVolumeSpecName "kube-api-access-tpdfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:18.425744 master-0 kubenswrapper[24928]: I1205 11:07:18.425667 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlzkr\" (UniqueName: \"kubernetes.io/projected/e73ed846-11a8-4c4b-b779-d7d362a3933f-kube-api-access-dlzkr\") pod \"e73ed846-11a8-4c4b-b779-d7d362a3933f\" (UID: \"e73ed846-11a8-4c4b-b779-d7d362a3933f\") " Dec 05 11:07:18.425744 master-0 kubenswrapper[24928]: I1205 11:07:18.425750 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82b65c14-8968-437d-840d-30ccd28c57f9-operator-scripts\") pod \"82b65c14-8968-437d-840d-30ccd28c57f9\" (UID: \"82b65c14-8968-437d-840d-30ccd28c57f9\") " Dec 05 11:07:18.426010 master-0 kubenswrapper[24928]: I1205 11:07:18.425875 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2vtl\" (UniqueName: \"kubernetes.io/projected/82b65c14-8968-437d-840d-30ccd28c57f9-kube-api-access-z2vtl\") pod \"82b65c14-8968-437d-840d-30ccd28c57f9\" (UID: \"82b65c14-8968-437d-840d-30ccd28c57f9\") " Dec 05 11:07:18.426059 master-0 kubenswrapper[24928]: I1205 11:07:18.426029 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e73ed846-11a8-4c4b-b779-d7d362a3933f-operator-scripts\") pod \"e73ed846-11a8-4c4b-b779-d7d362a3933f\" (UID: \"e73ed846-11a8-4c4b-b779-d7d362a3933f\") " Dec 05 11:07:18.426305 master-0 kubenswrapper[24928]: I1205 11:07:18.426258 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82b65c14-8968-437d-840d-30ccd28c57f9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "82b65c14-8968-437d-840d-30ccd28c57f9" (UID: "82b65c14-8968-437d-840d-30ccd28c57f9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:07:18.426575 master-0 kubenswrapper[24928]: I1205 11:07:18.426532 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e73ed846-11a8-4c4b-b779-d7d362a3933f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e73ed846-11a8-4c4b-b779-d7d362a3933f" (UID: "e73ed846-11a8-4c4b-b779-d7d362a3933f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:07:18.428741 master-0 kubenswrapper[24928]: I1205 11:07:18.428710 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e73ed846-11a8-4c4b-b779-d7d362a3933f-kube-api-access-dlzkr" (OuterVolumeSpecName: "kube-api-access-dlzkr") pod "e73ed846-11a8-4c4b-b779-d7d362a3933f" (UID: "e73ed846-11a8-4c4b-b779-d7d362a3933f"). InnerVolumeSpecName "kube-api-access-dlzkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:18.429332 master-0 kubenswrapper[24928]: I1205 11:07:18.429160 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82b65c14-8968-437d-840d-30ccd28c57f9-kube-api-access-z2vtl" (OuterVolumeSpecName: "kube-api-access-z2vtl") pod "82b65c14-8968-437d-840d-30ccd28c57f9" (UID: "82b65c14-8968-437d-840d-30ccd28c57f9"). InnerVolumeSpecName "kube-api-access-z2vtl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:18.429872 master-0 kubenswrapper[24928]: I1205 11:07:18.429844 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpdfv\" (UniqueName: \"kubernetes.io/projected/f180288f-009b-4f64-bf07-543726ebc2e2-kube-api-access-tpdfv\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:18.429872 master-0 kubenswrapper[24928]: I1205 11:07:18.429870 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f180288f-009b-4f64-bf07-543726ebc2e2-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:18.430004 master-0 kubenswrapper[24928]: I1205 11:07:18.429881 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlzkr\" (UniqueName: \"kubernetes.io/projected/e73ed846-11a8-4c4b-b779-d7d362a3933f-kube-api-access-dlzkr\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:18.430004 master-0 kubenswrapper[24928]: I1205 11:07:18.429892 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/82b65c14-8968-437d-840d-30ccd28c57f9-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:18.430004 master-0 kubenswrapper[24928]: I1205 11:07:18.429912 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2vtl\" (UniqueName: \"kubernetes.io/projected/82b65c14-8968-437d-840d-30ccd28c57f9-kube-api-access-z2vtl\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:18.430004 master-0 kubenswrapper[24928]: I1205 11:07:18.429922 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e73ed846-11a8-4c4b-b779-d7d362a3933f-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:18.572050 master-0 kubenswrapper[24928]: I1205 11:07:18.571986 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9b34-account-create-update-gllvm" Dec 05 11:07:18.572267 master-0 kubenswrapper[24928]: I1205 11:07:18.571997 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9b34-account-create-update-gllvm" event={"ID":"82b65c14-8968-437d-840d-30ccd28c57f9","Type":"ContainerDied","Data":"da5a0fa9aa6942984dd3451c39990e1cb3614bdaabb8f285637ac1a90ccbf20b"} Dec 05 11:07:18.572310 master-0 kubenswrapper[24928]: I1205 11:07:18.572244 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da5a0fa9aa6942984dd3451c39990e1cb3614bdaabb8f285637ac1a90ccbf20b" Dec 05 11:07:18.574634 master-0 kubenswrapper[24928]: I1205 11:07:18.574582 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-649rk" event={"ID":"e73ed846-11a8-4c4b-b779-d7d362a3933f","Type":"ContainerDied","Data":"a07d4f0dbb9d6e35e5c60d8f36d62c53539e2e1faf1b99fb77c1f9eac54c5527"} Dec 05 11:07:18.574711 master-0 kubenswrapper[24928]: I1205 11:07:18.574642 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a07d4f0dbb9d6e35e5c60d8f36d62c53539e2e1faf1b99fb77c1f9eac54c5527" Dec 05 11:07:18.574711 master-0 kubenswrapper[24928]: I1205 11:07:18.574667 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-649rk" Dec 05 11:07:18.577304 master-0 kubenswrapper[24928]: I1205 11:07:18.577275 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1c49-account-create-update-vwpvg" event={"ID":"f180288f-009b-4f64-bf07-543726ebc2e2","Type":"ContainerDied","Data":"c609a74895d38f9693fc867343f0c4d6fac9f3cb2a4e895602733af49f1981ec"} Dec 05 11:07:18.577798 master-0 kubenswrapper[24928]: I1205 11:07:18.577309 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c609a74895d38f9693fc867343f0c4d6fac9f3cb2a4e895602733af49f1981ec" Dec 05 11:07:18.577798 master-0 kubenswrapper[24928]: I1205 11:07:18.577395 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1c49-account-create-update-vwpvg" Dec 05 11:07:19.598652 master-0 kubenswrapper[24928]: I1205 11:07:19.598590 24928 generic.go:334] "Generic (PLEG): container finished" podID="83a55bec-0267-4de3-97e8-8705a6056565" containerID="9fb6867aa7cfa7925e04c045e72c2da1fe1c373a0b7241f16eb9a30adbcfcf11" exitCode=0 Dec 05 11:07:19.598652 master-0 kubenswrapper[24928]: I1205 11:07:19.598647 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" event={"ID":"83a55bec-0267-4de3-97e8-8705a6056565","Type":"ContainerDied","Data":"9fb6867aa7cfa7925e04c045e72c2da1fe1c373a0b7241f16eb9a30adbcfcf11"} Dec 05 11:07:20.615712 master-0 kubenswrapper[24928]: I1205 11:07:20.615636 24928 generic.go:334] "Generic (PLEG): container finished" podID="2d8079b6-0de1-4ffd-b008-9d74bf9673ae" containerID="372563fe24f0f6f0192af404a2a04904a2370849e44934fb43245471fcdb49ff" exitCode=0 Dec 05 11:07:20.615712 master-0 kubenswrapper[24928]: I1205 11:07:20.615688 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"2d8079b6-0de1-4ffd-b008-9d74bf9673ae","Type":"ContainerDied","Data":"372563fe24f0f6f0192af404a2a04904a2370849e44934fb43245471fcdb49ff"} Dec 05 11:07:20.617749 master-0 kubenswrapper[24928]: I1205 11:07:20.617717 24928 generic.go:334] "Generic (PLEG): container finished" podID="06c47656-d5b2-405d-944d-3d4ce03be7b0" containerID="4e24ad44d524026003f4888853d5d3caeac4e3f2ac890101297980e102d96974" exitCode=0 Dec 05 11:07:20.618142 master-0 kubenswrapper[24928]: I1205 11:07:20.617862 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g2gnb" event={"ID":"06c47656-d5b2-405d-944d-3d4ce03be7b0","Type":"ContainerDied","Data":"4e24ad44d524026003f4888853d5d3caeac4e3f2ac890101297980e102d96974"} Dec 05 11:07:20.622335 master-0 kubenswrapper[24928]: I1205 11:07:20.622283 24928 generic.go:334] "Generic (PLEG): container finished" podID="9c86a407-7dbd-49d1-87e8-c447bceb67bb" containerID="0b891267e1e08f06ff8f2233654172342f8078c9cf3a0e2fc711005a5b723ba8" exitCode=0 Dec 05 11:07:20.622521 master-0 kubenswrapper[24928]: I1205 11:07:20.622380 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tqgcv" event={"ID":"9c86a407-7dbd-49d1-87e8-c447bceb67bb","Type":"ContainerDied","Data":"0b891267e1e08f06ff8f2233654172342f8078c9cf3a0e2fc711005a5b723ba8"} Dec 05 11:07:20.625041 master-0 kubenswrapper[24928]: I1205 11:07:20.625006 24928 generic.go:334] "Generic (PLEG): container finished" podID="fddf0e6c-7977-4429-9015-732fd8dbb6a1" containerID="2130b735befea54638f55b346f56e9884c2c0ab93e77888bb7dad2d68c56bea0" exitCode=0 Dec 05 11:07:20.625193 master-0 kubenswrapper[24928]: I1205 11:07:20.625127 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"fddf0e6c-7977-4429-9015-732fd8dbb6a1","Type":"ContainerDied","Data":"2130b735befea54638f55b346f56e9884c2c0ab93e77888bb7dad2d68c56bea0"} Dec 05 11:07:21.180503 master-0 kubenswrapper[24928]: I1205 11:07:21.180415 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" Dec 05 11:07:21.423941 master-0 kubenswrapper[24928]: I1205 11:07:21.423809 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-neutron-agent-66b77f87b6-l2d4s" Dec 05 11:07:21.642924 master-0 kubenswrapper[24928]: I1205 11:07:21.642830 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" Dec 05 11:07:21.643703 master-0 kubenswrapper[24928]: I1205 11:07:21.643183 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2b0e-account-create-update-fpn9k" event={"ID":"83a55bec-0267-4de3-97e8-8705a6056565","Type":"ContainerDied","Data":"241e80b1b0c15d2da03ed575e6017698f627610a44995dbdc95e2691ff98624e"} Dec 05 11:07:21.643703 master-0 kubenswrapper[24928]: I1205 11:07:21.643259 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="241e80b1b0c15d2da03ed575e6017698f627610a44995dbdc95e2691ff98624e" Dec 05 11:07:21.645505 master-0 kubenswrapper[24928]: I1205 11:07:21.645447 24928 generic.go:334] "Generic (PLEG): container finished" podID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerID="acd53d82e3153e795b5309b7c86275854aea6e0fe04723a35820bec9073a5ad8" exitCode=0 Dec 05 11:07:21.645618 master-0 kubenswrapper[24928]: I1205 11:07:21.645512 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"e1a0529b-71c2-4a72-b75f-0b56b07e7c64","Type":"ContainerDied","Data":"acd53d82e3153e795b5309b7c86275854aea6e0fe04723a35820bec9073a5ad8"} Dec 05 11:07:21.647353 master-0 kubenswrapper[24928]: I1205 11:07:21.647304 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" event={"ID":"b7a38786-cb62-46b5-8a00-54f6d01545b8","Type":"ContainerStarted","Data":"6d1b122f166fe65e8e7588f3cd14d0233e19f369beb8056737fd38fc24cd8f3a"} Dec 05 11:07:21.648127 master-0 kubenswrapper[24928]: I1205 11:07:21.648084 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:21.649840 master-0 kubenswrapper[24928]: I1205 11:07:21.649809 24928 generic.go:334] "Generic (PLEG): container finished" podID="313a9b9c-4454-4927-a218-761e791cf156" containerID="18b57fa2300e0730e4f472988dc939571cd207bbee3ec3cc9df53f38fc2ea267" exitCode=0 Dec 05 11:07:21.665499 master-0 kubenswrapper[24928]: I1205 11:07:21.652371 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"313a9b9c-4454-4927-a218-761e791cf156","Type":"ContainerDied","Data":"18b57fa2300e0730e4f472988dc939571cd207bbee3ec3cc9df53f38fc2ea267"} Dec 05 11:07:22.239631 master-0 kubenswrapper[24928]: I1205 11:07:22.239568 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g2gnb" Dec 05 11:07:22.246592 master-0 kubenswrapper[24928]: I1205 11:07:22.246547 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tqgcv" Dec 05 11:07:22.703920 master-0 kubenswrapper[24928]: I1205 11:07:22.703861 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tqgcv" event={"ID":"9c86a407-7dbd-49d1-87e8-c447bceb67bb","Type":"ContainerDied","Data":"b706a0c33fc18345c5b0f3f1d8ed5d311a2781bad2dc17e66152d9f3811a8bc0"} Dec 05 11:07:22.703920 master-0 kubenswrapper[24928]: I1205 11:07:22.703924 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b706a0c33fc18345c5b0f3f1d8ed5d311a2781bad2dc17e66152d9f3811a8bc0" Dec 05 11:07:22.704496 master-0 kubenswrapper[24928]: I1205 11:07:22.704007 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tqgcv" Dec 05 11:07:22.750447 master-0 kubenswrapper[24928]: I1205 11:07:22.739570 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83a55bec-0267-4de3-97e8-8705a6056565-operator-scripts\") pod \"83a55bec-0267-4de3-97e8-8705a6056565\" (UID: \"83a55bec-0267-4de3-97e8-8705a6056565\") " Dec 05 11:07:22.750447 master-0 kubenswrapper[24928]: I1205 11:07:22.739654 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c86a407-7dbd-49d1-87e8-c447bceb67bb-operator-scripts\") pod \"9c86a407-7dbd-49d1-87e8-c447bceb67bb\" (UID: \"9c86a407-7dbd-49d1-87e8-c447bceb67bb\") " Dec 05 11:07:22.750447 master-0 kubenswrapper[24928]: I1205 11:07:22.739837 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsggq\" (UniqueName: \"kubernetes.io/projected/06c47656-d5b2-405d-944d-3d4ce03be7b0-kube-api-access-fsggq\") pod \"06c47656-d5b2-405d-944d-3d4ce03be7b0\" (UID: \"06c47656-d5b2-405d-944d-3d4ce03be7b0\") " Dec 05 11:07:22.750447 master-0 kubenswrapper[24928]: I1205 11:07:22.740017 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06c47656-d5b2-405d-944d-3d4ce03be7b0-operator-scripts\") pod \"06c47656-d5b2-405d-944d-3d4ce03be7b0\" (UID: \"06c47656-d5b2-405d-944d-3d4ce03be7b0\") " Dec 05 11:07:22.750447 master-0 kubenswrapper[24928]: I1205 11:07:22.740075 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfvlv\" (UniqueName: \"kubernetes.io/projected/9c86a407-7dbd-49d1-87e8-c447bceb67bb-kube-api-access-jfvlv\") pod \"9c86a407-7dbd-49d1-87e8-c447bceb67bb\" (UID: \"9c86a407-7dbd-49d1-87e8-c447bceb67bb\") " Dec 05 11:07:22.750447 master-0 kubenswrapper[24928]: I1205 11:07:22.740128 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mllp\" (UniqueName: \"kubernetes.io/projected/83a55bec-0267-4de3-97e8-8705a6056565-kube-api-access-4mllp\") pod \"83a55bec-0267-4de3-97e8-8705a6056565\" (UID: \"83a55bec-0267-4de3-97e8-8705a6056565\") " Dec 05 11:07:22.756451 master-0 kubenswrapper[24928]: I1205 11:07:22.751178 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a55bec-0267-4de3-97e8-8705a6056565-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "83a55bec-0267-4de3-97e8-8705a6056565" (UID: "83a55bec-0267-4de3-97e8-8705a6056565"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:07:22.756451 master-0 kubenswrapper[24928]: I1205 11:07:22.751696 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c86a407-7dbd-49d1-87e8-c447bceb67bb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c86a407-7dbd-49d1-87e8-c447bceb67bb" (UID: "9c86a407-7dbd-49d1-87e8-c447bceb67bb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:07:22.756451 master-0 kubenswrapper[24928]: I1205 11:07:22.752158 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06c47656-d5b2-405d-944d-3d4ce03be7b0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "06c47656-d5b2-405d-944d-3d4ce03be7b0" (UID: "06c47656-d5b2-405d-944d-3d4ce03be7b0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:07:22.762867 master-0 kubenswrapper[24928]: I1205 11:07:22.762782 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a55bec-0267-4de3-97e8-8705a6056565-kube-api-access-4mllp" (OuterVolumeSpecName: "kube-api-access-4mllp") pod "83a55bec-0267-4de3-97e8-8705a6056565" (UID: "83a55bec-0267-4de3-97e8-8705a6056565"). InnerVolumeSpecName "kube-api-access-4mllp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:22.781950 master-0 kubenswrapper[24928]: I1205 11:07:22.780906 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c86a407-7dbd-49d1-87e8-c447bceb67bb-kube-api-access-jfvlv" (OuterVolumeSpecName: "kube-api-access-jfvlv") pod "9c86a407-7dbd-49d1-87e8-c447bceb67bb" (UID: "9c86a407-7dbd-49d1-87e8-c447bceb67bb"). InnerVolumeSpecName "kube-api-access-jfvlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:22.781950 master-0 kubenswrapper[24928]: I1205 11:07:22.781042 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-g2gnb" Dec 05 11:07:22.781950 master-0 kubenswrapper[24928]: I1205 11:07:22.781068 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06c47656-d5b2-405d-944d-3d4ce03be7b0-kube-api-access-fsggq" (OuterVolumeSpecName: "kube-api-access-fsggq") pod "06c47656-d5b2-405d-944d-3d4ce03be7b0" (UID: "06c47656-d5b2-405d-944d-3d4ce03be7b0"). InnerVolumeSpecName "kube-api-access-fsggq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:22.781950 master-0 kubenswrapper[24928]: I1205 11:07:22.781083 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-g2gnb" event={"ID":"06c47656-d5b2-405d-944d-3d4ce03be7b0","Type":"ContainerDied","Data":"38cf6d9229224b5097ccca695b90e356df9d189e14979f2fca3757c20f7af0b2"} Dec 05 11:07:22.781950 master-0 kubenswrapper[24928]: I1205 11:07:22.781193 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38cf6d9229224b5097ccca695b90e356df9d189e14979f2fca3757c20f7af0b2" Dec 05 11:07:22.846500 master-0 kubenswrapper[24928]: I1205 11:07:22.843583 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsggq\" (UniqueName: \"kubernetes.io/projected/06c47656-d5b2-405d-944d-3d4ce03be7b0-kube-api-access-fsggq\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:22.846500 master-0 kubenswrapper[24928]: I1205 11:07:22.843641 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06c47656-d5b2-405d-944d-3d4ce03be7b0-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:22.846500 master-0 kubenswrapper[24928]: I1205 11:07:22.843657 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jfvlv\" (UniqueName: \"kubernetes.io/projected/9c86a407-7dbd-49d1-87e8-c447bceb67bb-kube-api-access-jfvlv\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:22.846500 master-0 kubenswrapper[24928]: I1205 11:07:22.843669 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mllp\" (UniqueName: \"kubernetes.io/projected/83a55bec-0267-4de3-97e8-8705a6056565-kube-api-access-4mllp\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:22.846500 master-0 kubenswrapper[24928]: I1205 11:07:22.843682 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/83a55bec-0267-4de3-97e8-8705a6056565-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:22.846500 master-0 kubenswrapper[24928]: I1205 11:07:22.843695 24928 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c86a407-7dbd-49d1-87e8-c447bceb67bb-operator-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:22.859723 master-0 kubenswrapper[24928]: I1205 11:07:22.849647 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-ec941-default-external-api-0" podUID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerName="glance-log" probeResult="failure" output="Get \"https://10.128.0.211:9292/healthcheck\": dial tcp 10.128.0.211:9292: connect: connection refused" Dec 05 11:07:22.859723 master-0 kubenswrapper[24928]: I1205 11:07:22.849952 24928 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-ec941-default-external-api-0" podUID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.128.0.211:9292/healthcheck\": dial tcp 10.128.0.211:9292: connect: connection refused" Dec 05 11:07:22.912869 master-0 kubenswrapper[24928]: I1205 11:07:22.897843 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" podStartSLOduration=8.897818788 podStartE2EDuration="8.897818788s" podCreationTimestamp="2025-12-05 11:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:07:22.871378463 +0000 UTC m=+1202.874572314" watchObservedRunningTime="2025-12-05 11:07:22.897818788 +0000 UTC m=+1202.901012639" Dec 05 11:07:23.312678 master-0 kubenswrapper[24928]: I1205 11:07:23.312600 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:23.478685 master-0 kubenswrapper[24928]: I1205 11:07:23.478393 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ktrt\" (UniqueName: \"kubernetes.io/projected/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-kube-api-access-6ktrt\") pod \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " Dec 05 11:07:23.479445 master-0 kubenswrapper[24928]: I1205 11:07:23.479315 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-scripts\") pod \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " Dec 05 11:07:23.480810 master-0 kubenswrapper[24928]: I1205 11:07:23.480678 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-combined-ca-bundle\") pod \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " Dec 05 11:07:23.482577 master-0 kubenswrapper[24928]: I1205 11:07:23.482541 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-config-data\") pod \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " Dec 05 11:07:23.482842 master-0 kubenswrapper[24928]: I1205 11:07:23.482787 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-kube-api-access-6ktrt" (OuterVolumeSpecName: "kube-api-access-6ktrt") pod "e1a0529b-71c2-4a72-b75f-0b56b07e7c64" (UID: "e1a0529b-71c2-4a72-b75f-0b56b07e7c64"). InnerVolumeSpecName "kube-api-access-6ktrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:23.484199 master-0 kubenswrapper[24928]: I1205 11:07:23.484164 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " Dec 05 11:07:23.484295 master-0 kubenswrapper[24928]: I1205 11:07:23.484209 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-public-tls-certs\") pod \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " Dec 05 11:07:23.484295 master-0 kubenswrapper[24928]: I1205 11:07:23.484250 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-logs\") pod \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " Dec 05 11:07:23.484377 master-0 kubenswrapper[24928]: I1205 11:07:23.484334 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-httpd-run\") pod \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\" (UID: \"e1a0529b-71c2-4a72-b75f-0b56b07e7c64\") " Dec 05 11:07:23.485785 master-0 kubenswrapper[24928]: I1205 11:07:23.485727 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-logs" (OuterVolumeSpecName: "logs") pod "e1a0529b-71c2-4a72-b75f-0b56b07e7c64" (UID: "e1a0529b-71c2-4a72-b75f-0b56b07e7c64"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:07:23.486372 master-0 kubenswrapper[24928]: I1205 11:07:23.486325 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e1a0529b-71c2-4a72-b75f-0b56b07e7c64" (UID: "e1a0529b-71c2-4a72-b75f-0b56b07e7c64"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:07:23.490315 master-0 kubenswrapper[24928]: I1205 11:07:23.490268 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ktrt\" (UniqueName: \"kubernetes.io/projected/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-kube-api-access-6ktrt\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:23.490390 master-0 kubenswrapper[24928]: I1205 11:07:23.490318 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:23.490390 master-0 kubenswrapper[24928]: I1205 11:07:23.490330 24928 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-httpd-run\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:23.490654 master-0 kubenswrapper[24928]: I1205 11:07:23.490410 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-scripts" (OuterVolumeSpecName: "scripts") pod "e1a0529b-71c2-4a72-b75f-0b56b07e7c64" (UID: "e1a0529b-71c2-4a72-b75f-0b56b07e7c64"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:23.509698 master-0 kubenswrapper[24928]: I1205 11:07:23.505081 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19" (OuterVolumeSpecName: "glance") pod "e1a0529b-71c2-4a72-b75f-0b56b07e7c64" (UID: "e1a0529b-71c2-4a72-b75f-0b56b07e7c64"). InnerVolumeSpecName "pvc-2a5fd065-f556-4411-b867-061babb4188e". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 11:07:23.527451 master-0 kubenswrapper[24928]: I1205 11:07:23.524984 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1a0529b-71c2-4a72-b75f-0b56b07e7c64" (UID: "e1a0529b-71c2-4a72-b75f-0b56b07e7c64"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:23.594106 master-0 kubenswrapper[24928]: I1205 11:07:23.592748 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:23.594106 master-0 kubenswrapper[24928]: I1205 11:07:23.592810 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:23.594106 master-0 kubenswrapper[24928]: I1205 11:07:23.592847 24928 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") on node \"master-0\" " Dec 05 11:07:23.594106 master-0 kubenswrapper[24928]: I1205 11:07:23.592909 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e1a0529b-71c2-4a72-b75f-0b56b07e7c64" (UID: "e1a0529b-71c2-4a72-b75f-0b56b07e7c64"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:23.644736 master-0 kubenswrapper[24928]: I1205 11:07:23.644672 24928 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 11:07:23.645002 master-0 kubenswrapper[24928]: I1205 11:07:23.644879 24928 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-2a5fd065-f556-4411-b867-061babb4188e" (UniqueName: "kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19") on node "master-0" Dec 05 11:07:23.659871 master-0 kubenswrapper[24928]: I1205 11:07:23.659694 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-config-data" (OuterVolumeSpecName: "config-data") pod "e1a0529b-71c2-4a72-b75f-0b56b07e7c64" (UID: "e1a0529b-71c2-4a72-b75f-0b56b07e7c64"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:23.702143 master-0 kubenswrapper[24928]: I1205 11:07:23.702054 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:23.702143 master-0 kubenswrapper[24928]: I1205 11:07:23.702133 24928 reconciler_common.go:293] "Volume detached for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:23.702143 master-0 kubenswrapper[24928]: I1205 11:07:23.702150 24928 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0529b-71c2-4a72-b75f-0b56b07e7c64-public-tls-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:23.805696 master-0 kubenswrapper[24928]: I1205 11:07:23.805619 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"e1a0529b-71c2-4a72-b75f-0b56b07e7c64","Type":"ContainerDied","Data":"e091c68b96876c484a694f5b55438623d568ae6bf609dd6717299b74ac97cc8e"} Dec 05 11:07:23.806208 master-0 kubenswrapper[24928]: I1205 11:07:23.805716 24928 scope.go:117] "RemoveContainer" containerID="acd53d82e3153e795b5309b7c86275854aea6e0fe04723a35820bec9073a5ad8" Dec 05 11:07:23.806208 master-0 kubenswrapper[24928]: I1205 11:07:23.805656 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:23.854435 master-0 kubenswrapper[24928]: I1205 11:07:23.853790 24928 scope.go:117] "RemoveContainer" containerID="73bb8e24a58ffbe2967e4ba93ebbac3bff4baac27f7243d2c461aaa0802ea5d2" Dec 05 11:07:23.906557 master-0 kubenswrapper[24928]: I1205 11:07:23.904877 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:23.913816 master-0 kubenswrapper[24928]: I1205 11:07:23.913650 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-internal-tls-certs\") pod \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " Dec 05 11:07:23.913816 master-0 kubenswrapper[24928]: I1205 11:07:23.913746 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-combined-ca-bundle\") pod \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " Dec 05 11:07:23.914102 master-0 kubenswrapper[24928]: I1205 11:07:23.913848 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-scripts\") pod \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " Dec 05 11:07:23.914102 master-0 kubenswrapper[24928]: I1205 11:07:23.913913 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-config-data\") pod \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " Dec 05 11:07:23.914668 master-0 kubenswrapper[24928]: I1205 11:07:23.914619 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-logs\") pod \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " Dec 05 11:07:23.915087 master-0 kubenswrapper[24928]: I1205 11:07:23.915052 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " Dec 05 11:07:23.915160 master-0 kubenswrapper[24928]: I1205 11:07:23.915118 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-httpd-run\") pod \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " Dec 05 11:07:23.915884 master-0 kubenswrapper[24928]: I1205 11:07:23.915847 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fddf0e6c-7977-4429-9015-732fd8dbb6a1" (UID: "fddf0e6c-7977-4429-9015-732fd8dbb6a1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:07:23.916183 master-0 kubenswrapper[24928]: I1205 11:07:23.916153 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-logs" (OuterVolumeSpecName: "logs") pod "fddf0e6c-7977-4429-9015-732fd8dbb6a1" (UID: "fddf0e6c-7977-4429-9015-732fd8dbb6a1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:07:23.918142 master-0 kubenswrapper[24928]: I1205 11:07:23.918064 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-scripts" (OuterVolumeSpecName: "scripts") pod "fddf0e6c-7977-4429-9015-732fd8dbb6a1" (UID: "fddf0e6c-7977-4429-9015-732fd8dbb6a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:23.919412 master-0 kubenswrapper[24928]: I1205 11:07:23.919234 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:23.919412 master-0 kubenswrapper[24928]: I1205 11:07:23.919385 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:23.919412 master-0 kubenswrapper[24928]: I1205 11:07:23.919399 24928 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fddf0e6c-7977-4429-9015-732fd8dbb6a1-httpd-run\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:23.921806 master-0 kubenswrapper[24928]: I1205 11:07:23.921751 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.961379 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.961485 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: E1205 11:07:23.961942 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f180288f-009b-4f64-bf07-543726ebc2e2" containerName="mariadb-account-create-update" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.961955 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f180288f-009b-4f64-bf07-543726ebc2e2" containerName="mariadb-account-create-update" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: E1205 11:07:23.961982 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06c47656-d5b2-405d-944d-3d4ce03be7b0" containerName="mariadb-database-create" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.961989 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="06c47656-d5b2-405d-944d-3d4ce03be7b0" containerName="mariadb-database-create" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: E1205 11:07:23.961999 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddf0e6c-7977-4429-9015-732fd8dbb6a1" containerName="glance-httpd" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962005 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddf0e6c-7977-4429-9015-732fd8dbb6a1" containerName="glance-httpd" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: E1205 11:07:23.962017 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a55bec-0267-4de3-97e8-8705a6056565" containerName="mariadb-account-create-update" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962024 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a55bec-0267-4de3-97e8-8705a6056565" containerName="mariadb-account-create-update" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: E1205 11:07:23.962042 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerName="glance-httpd" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962050 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerName="glance-httpd" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: E1205 11:07:23.962073 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerName="glance-log" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962078 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerName="glance-log" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: E1205 11:07:23.962100 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c86a407-7dbd-49d1-87e8-c447bceb67bb" containerName="mariadb-database-create" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962106 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c86a407-7dbd-49d1-87e8-c447bceb67bb" containerName="mariadb-database-create" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: E1205 11:07:23.962119 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b65c14-8968-437d-840d-30ccd28c57f9" containerName="mariadb-account-create-update" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962125 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b65c14-8968-437d-840d-30ccd28c57f9" containerName="mariadb-account-create-update" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: E1205 11:07:23.962138 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fddf0e6c-7977-4429-9015-732fd8dbb6a1" containerName="glance-log" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962144 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="fddf0e6c-7977-4429-9015-732fd8dbb6a1" containerName="glance-log" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: E1205 11:07:23.962155 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e73ed846-11a8-4c4b-b779-d7d362a3933f" containerName="mariadb-database-create" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962162 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="e73ed846-11a8-4c4b-b779-d7d362a3933f" containerName="mariadb-database-create" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962358 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerName="glance-httpd" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962377 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="fddf0e6c-7977-4429-9015-732fd8dbb6a1" containerName="glance-httpd" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962385 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="fddf0e6c-7977-4429-9015-732fd8dbb6a1" containerName="glance-log" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962444 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="e73ed846-11a8-4c4b-b779-d7d362a3933f" containerName="mariadb-database-create" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962457 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f180288f-009b-4f64-bf07-543726ebc2e2" containerName="mariadb-account-create-update" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962468 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a55bec-0267-4de3-97e8-8705a6056565" containerName="mariadb-account-create-update" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962483 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="06c47656-d5b2-405d-944d-3d4ce03be7b0" containerName="mariadb-database-create" Dec 05 11:07:23.963740 master-0 kubenswrapper[24928]: I1205 11:07:23.962506 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="82b65c14-8968-437d-840d-30ccd28c57f9" containerName="mariadb-account-create-update" Dec 05 11:07:23.965335 master-0 kubenswrapper[24928]: I1205 11:07:23.962518 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c86a407-7dbd-49d1-87e8-c447bceb67bb" containerName="mariadb-database-create" Dec 05 11:07:23.965335 master-0 kubenswrapper[24928]: I1205 11:07:23.964881 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" containerName="glance-log" Dec 05 11:07:23.967513 master-0 kubenswrapper[24928]: I1205 11:07:23.967234 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:23.971058 master-0 kubenswrapper[24928]: I1205 11:07:23.971015 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-ec941-default-external-config-data" Dec 05 11:07:23.971241 master-0 kubenswrapper[24928]: I1205 11:07:23.971180 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 11:07:23.971304 master-0 kubenswrapper[24928]: I1205 11:07:23.971228 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880" (OuterVolumeSpecName: "glance") pod "fddf0e6c-7977-4429-9015-732fd8dbb6a1" (UID: "fddf0e6c-7977-4429-9015-732fd8dbb6a1"). InnerVolumeSpecName "pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 11:07:23.985175 master-0 kubenswrapper[24928]: I1205 11:07:23.985106 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:07:23.987790 master-0 kubenswrapper[24928]: I1205 11:07:23.987741 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fddf0e6c-7977-4429-9015-732fd8dbb6a1" (UID: "fddf0e6c-7977-4429-9015-732fd8dbb6a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:24.022251 master-0 kubenswrapper[24928]: I1205 11:07:24.021048 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdjll\" (UniqueName: \"kubernetes.io/projected/fddf0e6c-7977-4429-9015-732fd8dbb6a1-kube-api-access-bdjll\") pod \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\" (UID: \"fddf0e6c-7977-4429-9015-732fd8dbb6a1\") " Dec 05 11:07:24.022552 master-0 kubenswrapper[24928]: I1205 11:07:24.022393 24928 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") on node \"master-0\" " Dec 05 11:07:24.022552 master-0 kubenswrapper[24928]: I1205 11:07:24.022436 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:24.033292 master-0 kubenswrapper[24928]: I1205 11:07:24.028254 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fddf0e6c-7977-4429-9015-732fd8dbb6a1-kube-api-access-bdjll" (OuterVolumeSpecName: "kube-api-access-bdjll") pod "fddf0e6c-7977-4429-9015-732fd8dbb6a1" (UID: "fddf0e6c-7977-4429-9015-732fd8dbb6a1"). InnerVolumeSpecName "kube-api-access-bdjll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:24.051600 master-0 kubenswrapper[24928]: I1205 11:07:24.051539 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fddf0e6c-7977-4429-9015-732fd8dbb6a1" (UID: "fddf0e6c-7977-4429-9015-732fd8dbb6a1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:24.057593 master-0 kubenswrapper[24928]: I1205 11:07:24.057524 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-config-data" (OuterVolumeSpecName: "config-data") pod "fddf0e6c-7977-4429-9015-732fd8dbb6a1" (UID: "fddf0e6c-7977-4429-9015-732fd8dbb6a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:24.072499 master-0 kubenswrapper[24928]: I1205 11:07:24.071203 24928 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 05 11:07:24.072499 master-0 kubenswrapper[24928]: I1205 11:07:24.071403 24928 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01" (UniqueName: "kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880") on node "master-0" Dec 05 11:07:24.124135 master-0 kubenswrapper[24928]: I1205 11:07:24.124047 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-httpd-run\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.124373 master-0 kubenswrapper[24928]: I1205 11:07:24.124161 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-logs\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.124373 master-0 kubenswrapper[24928]: I1205 11:07:24.124203 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-scripts\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.124373 master-0 kubenswrapper[24928]: I1205 11:07:24.124245 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-public-tls-certs\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.124373 master-0 kubenswrapper[24928]: I1205 11:07:24.124293 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zzmx\" (UniqueName: \"kubernetes.io/projected/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-kube-api-access-4zzmx\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.124373 master-0 kubenswrapper[24928]: I1205 11:07:24.124330 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-config-data\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.124729 master-0 kubenswrapper[24928]: I1205 11:07:24.124461 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-combined-ca-bundle\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.124729 master-0 kubenswrapper[24928]: I1205 11:07:24.124513 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.124729 master-0 kubenswrapper[24928]: I1205 11:07:24.124595 24928 reconciler_common.go:293] "Volume detached for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:24.124729 master-0 kubenswrapper[24928]: I1205 11:07:24.124615 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdjll\" (UniqueName: \"kubernetes.io/projected/fddf0e6c-7977-4429-9015-732fd8dbb6a1-kube-api-access-bdjll\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:24.124729 master-0 kubenswrapper[24928]: I1205 11:07:24.124630 24928 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-internal-tls-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:24.124729 master-0 kubenswrapper[24928]: I1205 11:07:24.124644 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fddf0e6c-7977-4429-9015-732fd8dbb6a1-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:24.223022 master-0 kubenswrapper[24928]: I1205 11:07:24.222730 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1a0529b-71c2-4a72-b75f-0b56b07e7c64" path="/var/lib/kubelet/pods/e1a0529b-71c2-4a72-b75f-0b56b07e7c64/volumes" Dec 05 11:07:24.227451 master-0 kubenswrapper[24928]: I1205 11:07:24.227018 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-httpd-run\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.227451 master-0 kubenswrapper[24928]: I1205 11:07:24.227120 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-logs\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.227451 master-0 kubenswrapper[24928]: I1205 11:07:24.227163 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-scripts\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.227451 master-0 kubenswrapper[24928]: I1205 11:07:24.227208 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-public-tls-certs\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.227451 master-0 kubenswrapper[24928]: I1205 11:07:24.227260 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zzmx\" (UniqueName: \"kubernetes.io/projected/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-kube-api-access-4zzmx\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.227451 master-0 kubenswrapper[24928]: I1205 11:07:24.227304 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-config-data\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.227451 master-0 kubenswrapper[24928]: I1205 11:07:24.227450 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-combined-ca-bundle\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.227947 master-0 kubenswrapper[24928]: I1205 11:07:24.227509 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.227947 master-0 kubenswrapper[24928]: I1205 11:07:24.227757 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-httpd-run\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.228754 master-0 kubenswrapper[24928]: I1205 11:07:24.228339 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-logs\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.231761 master-0 kubenswrapper[24928]: I1205 11:07:24.231082 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:07:24.231761 master-0 kubenswrapper[24928]: I1205 11:07:24.231144 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/4498c228462d6dd9d5d70d1a4aafdb52e778f9b7d49e65869f766f31c3adb188/globalmount\"" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.233769 master-0 kubenswrapper[24928]: I1205 11:07:24.232963 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-config-data\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.233860 master-0 kubenswrapper[24928]: I1205 11:07:24.233809 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-combined-ca-bundle\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.234652 master-0 kubenswrapper[24928]: I1205 11:07:24.234613 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-public-tls-certs\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.236451 master-0 kubenswrapper[24928]: I1205 11:07:24.236409 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-scripts\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.250062 master-0 kubenswrapper[24928]: I1205 11:07:24.250016 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zzmx\" (UniqueName: \"kubernetes.io/projected/d421cbb9-3e06-41b2-84ba-81ac38f1b07e-kube-api-access-4zzmx\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:24.840680 master-0 kubenswrapper[24928]: I1205 11:07:24.840599 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"fddf0e6c-7977-4429-9015-732fd8dbb6a1","Type":"ContainerDied","Data":"c42e7f420bef23b4105b5e3168b8c6a245f23a04964bd40194b0eb4e58b5bba3"} Dec 05 11:07:24.840680 master-0 kubenswrapper[24928]: I1205 11:07:24.840681 24928 scope.go:117] "RemoveContainer" containerID="2130b735befea54638f55b346f56e9884c2c0ab93e77888bb7dad2d68c56bea0" Dec 05 11:07:24.841304 master-0 kubenswrapper[24928]: I1205 11:07:24.840813 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:24.908395 master-0 kubenswrapper[24928]: I1205 11:07:24.908310 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:07:24.935618 master-0 kubenswrapper[24928]: I1205 11:07:24.935489 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:07:24.965811 master-0 kubenswrapper[24928]: I1205 11:07:24.965736 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:07:24.969092 master-0 kubenswrapper[24928]: I1205 11:07:24.968929 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:24.977783 master-0 kubenswrapper[24928]: I1205 11:07:24.977729 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 11:07:24.978074 master-0 kubenswrapper[24928]: I1205 11:07:24.978055 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-ec941-default-internal-config-data" Dec 05 11:07:24.978709 master-0 kubenswrapper[24928]: I1205 11:07:24.978688 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Dec 05 11:07:24.988766 master-0 kubenswrapper[24928]: I1205 11:07:24.988679 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:07:25.054054 master-0 kubenswrapper[24928]: I1205 11:07:25.053942 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1458275a-4e99-4257-88e5-e2ef36469ffc-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.054054 master-0 kubenswrapper[24928]: I1205 11:07:25.054008 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.054479 master-0 kubenswrapper[24928]: I1205 11:07:25.054138 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1458275a-4e99-4257-88e5-e2ef36469ffc-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.054479 master-0 kubenswrapper[24928]: I1205 11:07:25.054255 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-internal-tls-certs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.054548 master-0 kubenswrapper[24928]: I1205 11:07:25.054513 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9ghq\" (UniqueName: \"kubernetes.io/projected/1458275a-4e99-4257-88e5-e2ef36469ffc-kube-api-access-k9ghq\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.056613 master-0 kubenswrapper[24928]: I1205 11:07:25.054676 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.056613 master-0 kubenswrapper[24928]: I1205 11:07:25.054816 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.056613 master-0 kubenswrapper[24928]: I1205 11:07:25.054907 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.123945 master-0 kubenswrapper[24928]: I1205 11:07:25.123795 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2a5fd065-f556-4411-b867-061babb4188e\" (UniqueName: \"kubernetes.io/csi/topolvm.io^9b49b968-34a6-4357-b416-c76e6fb74d19\") pod \"glance-ec941-default-external-api-0\" (UID: \"d421cbb9-3e06-41b2-84ba-81ac38f1b07e\") " pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:25.158183 master-0 kubenswrapper[24928]: I1205 11:07:25.157234 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9ghq\" (UniqueName: \"kubernetes.io/projected/1458275a-4e99-4257-88e5-e2ef36469ffc-kube-api-access-k9ghq\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.158183 master-0 kubenswrapper[24928]: I1205 11:07:25.157336 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.158183 master-0 kubenswrapper[24928]: I1205 11:07:25.157394 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.158183 master-0 kubenswrapper[24928]: I1205 11:07:25.157495 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.158183 master-0 kubenswrapper[24928]: I1205 11:07:25.157529 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1458275a-4e99-4257-88e5-e2ef36469ffc-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.158183 master-0 kubenswrapper[24928]: I1205 11:07:25.157547 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.158183 master-0 kubenswrapper[24928]: I1205 11:07:25.157578 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1458275a-4e99-4257-88e5-e2ef36469ffc-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.158183 master-0 kubenswrapper[24928]: I1205 11:07:25.157617 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-internal-tls-certs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.160653 master-0 kubenswrapper[24928]: I1205 11:07:25.159706 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1458275a-4e99-4257-88e5-e2ef36469ffc-httpd-run\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.161465 master-0 kubenswrapper[24928]: I1205 11:07:25.160914 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1458275a-4e99-4257-88e5-e2ef36469ffc-logs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.162316 master-0 kubenswrapper[24928]: I1205 11:07:25.162279 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-internal-tls-certs\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.162508 master-0 kubenswrapper[24928]: I1205 11:07:25.162336 24928 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 11:07:25.162508 master-0 kubenswrapper[24928]: I1205 11:07:25.162367 24928 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/topolvm.io/f829e9071eee306944ccf41eabbcf74a1473d7031ba724c821f61b9a30a289e9/globalmount\"" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.164536 master-0 kubenswrapper[24928]: I1205 11:07:25.164163 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-scripts\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.169007 master-0 kubenswrapper[24928]: I1205 11:07:25.168443 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-combined-ca-bundle\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.174107 master-0 kubenswrapper[24928]: I1205 11:07:25.174058 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1458275a-4e99-4257-88e5-e2ef36469ffc-config-data\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.179074 master-0 kubenswrapper[24928]: I1205 11:07:25.179031 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9ghq\" (UniqueName: \"kubernetes.io/projected/1458275a-4e99-4257-88e5-e2ef36469ffc-kube-api-access-k9ghq\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:25.232485 master-0 kubenswrapper[24928]: I1205 11:07:25.232343 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:26.270175 master-0 kubenswrapper[24928]: I1205 11:07:26.270034 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fddf0e6c-7977-4429-9015-732fd8dbb6a1" path="/var/lib/kubelet/pods/fddf0e6c-7977-4429-9015-732fd8dbb6a1/volumes" Dec 05 11:07:26.879804 master-0 kubenswrapper[24928]: I1205 11:07:26.879364 24928 scope.go:117] "RemoveContainer" containerID="7c116d246b2d3f460a0aa28456f2c1fe5c53a3c50f6ed7a5a3f758d1ad2cb7c5" Dec 05 11:07:27.002158 master-0 kubenswrapper[24928]: I1205 11:07:26.999080 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-transport-url-ironic-inspector-transport" Dec 05 11:07:27.539970 master-0 kubenswrapper[24928]: I1205 11:07:27.539891 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec941-default-external-api-0"] Dec 05 11:07:27.897506 master-0 kubenswrapper[24928]: I1205 11:07:27.897209 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"2d8079b6-0de1-4ffd-b008-9d74bf9673ae","Type":"ContainerStarted","Data":"92d2c945339184594773d810897a8f3f1eab7fe95efacf43537287c8391cd24a"} Dec 05 11:07:27.902459 master-0 kubenswrapper[24928]: I1205 11:07:27.902391 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"313a9b9c-4454-4927-a218-761e791cf156","Type":"ContainerStarted","Data":"575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820"} Dec 05 11:07:27.902737 master-0 kubenswrapper[24928]: I1205 11:07:27.902640 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ironic-inspector-0" podUID="313a9b9c-4454-4927-a218-761e791cf156" containerName="inspector-pxe-init" containerID="cri-o://575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820" gracePeriod=60 Dec 05 11:07:27.904447 master-0 kubenswrapper[24928]: I1205 11:07:27.904261 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"d421cbb9-3e06-41b2-84ba-81ac38f1b07e","Type":"ContainerStarted","Data":"820b60c00744bb33f8cec1dcd5bff38d11da190638d50ad6d800de6086319691"} Dec 05 11:07:28.195866 master-0 kubenswrapper[24928]: I1205 11:07:28.195795 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fjthc"] Dec 05 11:07:28.197748 master-0 kubenswrapper[24928]: I1205 11:07:28.197517 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.201809 master-0 kubenswrapper[24928]: I1205 11:07:28.201767 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 05 11:07:28.201971 master-0 kubenswrapper[24928]: I1205 11:07:28.201766 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 11:07:28.236476 master-0 kubenswrapper[24928]: I1205 11:07:28.233323 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fjthc"] Dec 05 11:07:28.308405 master-0 kubenswrapper[24928]: I1205 11:07:28.308357 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-config-data\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.310009 master-0 kubenswrapper[24928]: I1205 11:07:28.309972 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-scripts\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.310113 master-0 kubenswrapper[24928]: I1205 11:07:28.310068 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4clb\" (UniqueName: \"kubernetes.io/projected/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-kube-api-access-t4clb\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.310505 master-0 kubenswrapper[24928]: I1205 11:07:28.310469 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.346623 master-0 kubenswrapper[24928]: I1205 11:07:28.346556 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-aa4ede8b-f742-4c34-84b7-5a8f3e754e01\" (UniqueName: \"kubernetes.io/csi/topolvm.io^870e5077-c116-4a47-9605-65093a1d6880\") pod \"glance-ec941-default-internal-api-0\" (UID: \"1458275a-4e99-4257-88e5-e2ef36469ffc\") " pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:28.415452 master-0 kubenswrapper[24928]: I1205 11:07:28.414451 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-scripts\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.415452 master-0 kubenswrapper[24928]: I1205 11:07:28.414600 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4clb\" (UniqueName: \"kubernetes.io/projected/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-kube-api-access-t4clb\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.415452 master-0 kubenswrapper[24928]: I1205 11:07:28.414840 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.415452 master-0 kubenswrapper[24928]: I1205 11:07:28.414952 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-config-data\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.427736 master-0 kubenswrapper[24928]: I1205 11:07:28.422980 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-config-data\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.450810 master-0 kubenswrapper[24928]: I1205 11:07:28.445056 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-scripts\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.473447 master-0 kubenswrapper[24928]: I1205 11:07:28.463124 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4clb\" (UniqueName: \"kubernetes.io/projected/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-kube-api-access-t4clb\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.496467 master-0 kubenswrapper[24928]: I1205 11:07:28.486672 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-fjthc\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.536448 master-0 kubenswrapper[24928]: I1205 11:07:28.536308 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:07:28.606437 master-0 kubenswrapper[24928]: I1205 11:07:28.606362 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:28.614298 master-0 kubenswrapper[24928]: I1205 11:07:28.614250 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Dec 05 11:07:28.730459 master-0 kubenswrapper[24928]: I1205 11:07:28.727191 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"313a9b9c-4454-4927-a218-761e791cf156\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " Dec 05 11:07:28.730459 master-0 kubenswrapper[24928]: I1205 11:07:28.727272 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic\") pod \"313a9b9c-4454-4927-a218-761e791cf156\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " Dec 05 11:07:28.730459 master-0 kubenswrapper[24928]: I1205 11:07:28.727313 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm5lr\" (UniqueName: \"kubernetes.io/projected/313a9b9c-4454-4927-a218-761e791cf156-kube-api-access-pm5lr\") pod \"313a9b9c-4454-4927-a218-761e791cf156\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " Dec 05 11:07:28.730459 master-0 kubenswrapper[24928]: I1205 11:07:28.727361 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-combined-ca-bundle\") pod \"313a9b9c-4454-4927-a218-761e791cf156\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " Dec 05 11:07:28.730459 master-0 kubenswrapper[24928]: I1205 11:07:28.727501 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-scripts\") pod \"313a9b9c-4454-4927-a218-761e791cf156\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " Dec 05 11:07:28.730459 master-0 kubenswrapper[24928]: I1205 11:07:28.727587 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-config\") pod \"313a9b9c-4454-4927-a218-761e791cf156\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " Dec 05 11:07:28.730459 master-0 kubenswrapper[24928]: I1205 11:07:28.727703 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/313a9b9c-4454-4927-a218-761e791cf156-etc-podinfo\") pod \"313a9b9c-4454-4927-a218-761e791cf156\" (UID: \"313a9b9c-4454-4927-a218-761e791cf156\") " Dec 05 11:07:28.730459 master-0 kubenswrapper[24928]: I1205 11:07:28.727802 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic-inspector-dhcp-hostsdir" (OuterVolumeSpecName: "var-lib-ironic-inspector-dhcp-hostsdir") pod "313a9b9c-4454-4927-a218-761e791cf156" (UID: "313a9b9c-4454-4927-a218-761e791cf156"). InnerVolumeSpecName "var-lib-ironic-inspector-dhcp-hostsdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:07:28.730459 master-0 kubenswrapper[24928]: I1205 11:07:28.728365 24928 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic-inspector-dhcp-hostsdir\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:28.732288 master-0 kubenswrapper[24928]: I1205 11:07:28.732214 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-config" (OuterVolumeSpecName: "config") pod "313a9b9c-4454-4927-a218-761e791cf156" (UID: "313a9b9c-4454-4927-a218-761e791cf156"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:28.733007 master-0 kubenswrapper[24928]: I1205 11:07:28.732936 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/313a9b9c-4454-4927-a218-761e791cf156-kube-api-access-pm5lr" (OuterVolumeSpecName: "kube-api-access-pm5lr") pod "313a9b9c-4454-4927-a218-761e791cf156" (UID: "313a9b9c-4454-4927-a218-761e791cf156"). InnerVolumeSpecName "kube-api-access-pm5lr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:28.734252 master-0 kubenswrapper[24928]: I1205 11:07:28.734127 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic" (OuterVolumeSpecName: "var-lib-ironic") pod "313a9b9c-4454-4927-a218-761e791cf156" (UID: "313a9b9c-4454-4927-a218-761e791cf156"). InnerVolumeSpecName "var-lib-ironic". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:07:28.738848 master-0 kubenswrapper[24928]: I1205 11:07:28.738770 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-scripts" (OuterVolumeSpecName: "scripts") pod "313a9b9c-4454-4927-a218-761e791cf156" (UID: "313a9b9c-4454-4927-a218-761e791cf156"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:28.749748 master-0 kubenswrapper[24928]: I1205 11:07:28.749666 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/313a9b9c-4454-4927-a218-761e791cf156-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "313a9b9c-4454-4927-a218-761e791cf156" (UID: "313a9b9c-4454-4927-a218-761e791cf156"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 11:07:28.832124 master-0 kubenswrapper[24928]: I1205 11:07:28.830538 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm5lr\" (UniqueName: \"kubernetes.io/projected/313a9b9c-4454-4927-a218-761e791cf156-kube-api-access-pm5lr\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:28.832124 master-0 kubenswrapper[24928]: I1205 11:07:28.830584 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:28.832124 master-0 kubenswrapper[24928]: I1205 11:07:28.830594 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:28.832124 master-0 kubenswrapper[24928]: I1205 11:07:28.830603 24928 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/313a9b9c-4454-4927-a218-761e791cf156-etc-podinfo\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:28.832124 master-0 kubenswrapper[24928]: I1205 11:07:28.830612 24928 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/313a9b9c-4454-4927-a218-761e791cf156-var-lib-ironic\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:28.832124 master-0 kubenswrapper[24928]: I1205 11:07:28.831651 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "313a9b9c-4454-4927-a218-761e791cf156" (UID: "313a9b9c-4454-4927-a218-761e791cf156"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:07:28.928834 master-0 kubenswrapper[24928]: I1205 11:07:28.928759 24928 generic.go:334] "Generic (PLEG): container finished" podID="313a9b9c-4454-4927-a218-761e791cf156" containerID="575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820" exitCode=0 Dec 05 11:07:28.929089 master-0 kubenswrapper[24928]: I1205 11:07:28.928864 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"313a9b9c-4454-4927-a218-761e791cf156","Type":"ContainerDied","Data":"575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820"} Dec 05 11:07:28.929089 master-0 kubenswrapper[24928]: I1205 11:07:28.928905 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"313a9b9c-4454-4927-a218-761e791cf156","Type":"ContainerDied","Data":"f12070fccf429dd91d728ac7daffd4a8fbe51645b1b9b50742a02ae76562f9ba"} Dec 05 11:07:28.929089 master-0 kubenswrapper[24928]: I1205 11:07:28.928926 24928 scope.go:117] "RemoveContainer" containerID="575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820" Dec 05 11:07:28.929089 master-0 kubenswrapper[24928]: I1205 11:07:28.929066 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Dec 05 11:07:28.944305 master-0 kubenswrapper[24928]: I1205 11:07:28.944252 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313a9b9c-4454-4927-a218-761e791cf156-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:28.995006 master-0 kubenswrapper[24928]: I1205 11:07:28.994951 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"d421cbb9-3e06-41b2-84ba-81ac38f1b07e","Type":"ContainerStarted","Data":"f474e03fcd314cb09178be57b3be817bc934d6b4960948f208e52f2306d95b78"} Dec 05 11:07:29.068626 master-0 kubenswrapper[24928]: I1205 11:07:29.066189 24928 scope.go:117] "RemoveContainer" containerID="18b57fa2300e0730e4f472988dc939571cd207bbee3ec3cc9df53f38fc2ea267" Dec 05 11:07:29.118274 master-0 kubenswrapper[24928]: I1205 11:07:29.105166 24928 scope.go:117] "RemoveContainer" containerID="575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820" Dec 05 11:07:29.140458 master-0 kubenswrapper[24928]: E1205 11:07:29.134770 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820\": container with ID starting with 575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820 not found: ID does not exist" containerID="575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820" Dec 05 11:07:29.140458 master-0 kubenswrapper[24928]: I1205 11:07:29.134848 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820"} err="failed to get container status \"575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820\": rpc error: code = NotFound desc = could not find container \"575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820\": container with ID starting with 575bad1b9b4f536af424e3ec1c55089acd8da002ef1c357e0683374323e0a820 not found: ID does not exist" Dec 05 11:07:29.140458 master-0 kubenswrapper[24928]: I1205 11:07:29.134875 24928 scope.go:117] "RemoveContainer" containerID="18b57fa2300e0730e4f472988dc939571cd207bbee3ec3cc9df53f38fc2ea267" Dec 05 11:07:29.140898 master-0 kubenswrapper[24928]: E1205 11:07:29.140840 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18b57fa2300e0730e4f472988dc939571cd207bbee3ec3cc9df53f38fc2ea267\": container with ID starting with 18b57fa2300e0730e4f472988dc939571cd207bbee3ec3cc9df53f38fc2ea267 not found: ID does not exist" containerID="18b57fa2300e0730e4f472988dc939571cd207bbee3ec3cc9df53f38fc2ea267" Dec 05 11:07:29.140964 master-0 kubenswrapper[24928]: I1205 11:07:29.140895 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18b57fa2300e0730e4f472988dc939571cd207bbee3ec3cc9df53f38fc2ea267"} err="failed to get container status \"18b57fa2300e0730e4f472988dc939571cd207bbee3ec3cc9df53f38fc2ea267\": rpc error: code = NotFound desc = could not find container \"18b57fa2300e0730e4f472988dc939571cd207bbee3ec3cc9df53f38fc2ea267\": container with ID starting with 18b57fa2300e0730e4f472988dc939571cd207bbee3ec3cc9df53f38fc2ea267 not found: ID does not exist" Dec 05 11:07:29.165711 master-0 kubenswrapper[24928]: W1205 11:07:29.165610 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b68cc45_2d4a_4377_ba5c_e35ecd4ed578.slice/crio-e7eb47c961ccc3f646a574f2974202d7027782a909645c7173030e1d846662b1 WatchSource:0}: Error finding container e7eb47c961ccc3f646a574f2974202d7027782a909645c7173030e1d846662b1: Status 404 returned error can't find the container with id e7eb47c961ccc3f646a574f2974202d7027782a909645c7173030e1d846662b1 Dec 05 11:07:29.192403 master-0 kubenswrapper[24928]: I1205 11:07:29.192337 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Dec 05 11:07:29.210512 master-0 kubenswrapper[24928]: I1205 11:07:29.208343 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fjthc"] Dec 05 11:07:29.282632 master-0 kubenswrapper[24928]: I1205 11:07:29.279617 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-0"] Dec 05 11:07:29.319553 master-0 kubenswrapper[24928]: I1205 11:07:29.316270 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-0"] Dec 05 11:07:29.319553 master-0 kubenswrapper[24928]: E1205 11:07:29.317046 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="313a9b9c-4454-4927-a218-761e791cf156" containerName="inspector-pxe-init" Dec 05 11:07:29.319553 master-0 kubenswrapper[24928]: I1205 11:07:29.317064 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="313a9b9c-4454-4927-a218-761e791cf156" containerName="inspector-pxe-init" Dec 05 11:07:29.319553 master-0 kubenswrapper[24928]: E1205 11:07:29.317094 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="313a9b9c-4454-4927-a218-761e791cf156" containerName="ironic-python-agent-init" Dec 05 11:07:29.319553 master-0 kubenswrapper[24928]: I1205 11:07:29.317100 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="313a9b9c-4454-4927-a218-761e791cf156" containerName="ironic-python-agent-init" Dec 05 11:07:29.319553 master-0 kubenswrapper[24928]: I1205 11:07:29.317363 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="313a9b9c-4454-4927-a218-761e791cf156" containerName="inspector-pxe-init" Dec 05 11:07:29.322448 master-0 kubenswrapper[24928]: I1205 11:07:29.322382 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Dec 05 11:07:29.329479 master-0 kubenswrapper[24928]: I1205 11:07:29.329199 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-transport-url-ironic-inspector-transport" Dec 05 11:07:29.329972 master-0 kubenswrapper[24928]: I1205 11:07:29.329722 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Dec 05 11:07:29.329972 master-0 kubenswrapper[24928]: I1205 11:07:29.329276 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-inspector-internal-svc" Dec 05 11:07:29.329972 master-0 kubenswrapper[24928]: I1205 11:07:29.329558 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Dec 05 11:07:29.330541 master-0 kubenswrapper[24928]: I1205 11:07:29.330156 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-inspector-public-svc" Dec 05 11:07:29.340793 master-0 kubenswrapper[24928]: I1205 11:07:29.339730 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Dec 05 11:07:29.363460 master-0 kubenswrapper[24928]: I1205 11:07:29.361626 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.363460 master-0 kubenswrapper[24928]: I1205 11:07:29.361778 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl658\" (UniqueName: \"kubernetes.io/projected/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-kube-api-access-wl658\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.363460 master-0 kubenswrapper[24928]: I1205 11:07:29.361817 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.363460 master-0 kubenswrapper[24928]: I1205 11:07:29.361862 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.363460 master-0 kubenswrapper[24928]: I1205 11:07:29.361928 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.363460 master-0 kubenswrapper[24928]: I1205 11:07:29.361960 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-config\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.363460 master-0 kubenswrapper[24928]: I1205 11:07:29.361987 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.363460 master-0 kubenswrapper[24928]: I1205 11:07:29.362044 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-scripts\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.363460 master-0 kubenswrapper[24928]: I1205 11:07:29.362092 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.382235 master-0 kubenswrapper[24928]: I1205 11:07:29.382164 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-ec941-default-internal-api-0"] Dec 05 11:07:29.393175 master-0 kubenswrapper[24928]: W1205 11:07:29.393104 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1458275a_4e99_4257_88e5_e2ef36469ffc.slice/crio-632cc173786d75b7b113f8ce9f24a3319e993801fb319dc0bd59c0f813e037c6 WatchSource:0}: Error finding container 632cc173786d75b7b113f8ce9f24a3319e993801fb319dc0bd59c0f813e037c6: Status 404 returned error can't find the container with id 632cc173786d75b7b113f8ce9f24a3319e993801fb319dc0bd59c0f813e037c6 Dec 05 11:07:29.464381 master-0 kubenswrapper[24928]: I1205 11:07:29.464296 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-scripts\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.465565 master-0 kubenswrapper[24928]: I1205 11:07:29.464727 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.465565 master-0 kubenswrapper[24928]: I1205 11:07:29.464863 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.465565 master-0 kubenswrapper[24928]: I1205 11:07:29.465048 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl658\" (UniqueName: \"kubernetes.io/projected/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-kube-api-access-wl658\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.465565 master-0 kubenswrapper[24928]: I1205 11:07:29.465106 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.465565 master-0 kubenswrapper[24928]: I1205 11:07:29.465161 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.465565 master-0 kubenswrapper[24928]: I1205 11:07:29.465182 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.466032 master-0 kubenswrapper[24928]: I1205 11:07:29.465624 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.466032 master-0 kubenswrapper[24928]: I1205 11:07:29.465679 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-config\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.466032 master-0 kubenswrapper[24928]: I1205 11:07:29.465711 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.466211 master-0 kubenswrapper[24928]: I1205 11:07:29.466166 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.469891 master-0 kubenswrapper[24928]: I1205 11:07:29.469822 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.470036 master-0 kubenswrapper[24928]: I1205 11:07:29.469898 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.470036 master-0 kubenswrapper[24928]: I1205 11:07:29.469845 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-scripts\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.471843 master-0 kubenswrapper[24928]: I1205 11:07:29.470452 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.472516 master-0 kubenswrapper[24928]: I1205 11:07:29.472273 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-config\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.472516 master-0 kubenswrapper[24928]: I1205 11:07:29.472503 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.512002 master-0 kubenswrapper[24928]: I1205 11:07:29.511869 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl658\" (UniqueName: \"kubernetes.io/projected/a44ef9f2-4b09-4875-b3d3-4ff3508980a1-kube-api-access-wl658\") pod \"ironic-inspector-0\" (UID: \"a44ef9f2-4b09-4875-b3d3-4ff3508980a1\") " pod="openstack/ironic-inspector-0" Dec 05 11:07:29.581535 master-0 kubenswrapper[24928]: I1205 11:07:29.576332 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:07:29.644025 master-0 kubenswrapper[24928]: I1205 11:07:29.643941 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Dec 05 11:07:30.017366 master-0 kubenswrapper[24928]: I1205 11:07:30.016465 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fjthc" event={"ID":"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578","Type":"ContainerStarted","Data":"e7eb47c961ccc3f646a574f2974202d7027782a909645c7173030e1d846662b1"} Dec 05 11:07:30.018514 master-0 kubenswrapper[24928]: I1205 11:07:30.018473 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"1458275a-4e99-4257-88e5-e2ef36469ffc","Type":"ContainerStarted","Data":"632cc173786d75b7b113f8ce9f24a3319e993801fb319dc0bd59c0f813e037c6"} Dec 05 11:07:30.022494 master-0 kubenswrapper[24928]: I1205 11:07:30.021911 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-external-api-0" event={"ID":"d421cbb9-3e06-41b2-84ba-81ac38f1b07e","Type":"ContainerStarted","Data":"9dd84f0c6237aca3e64d10464000ef54e0635f9328d362f4e6a2885d539bf3a0"} Dec 05 11:07:30.218325 master-0 kubenswrapper[24928]: I1205 11:07:30.218252 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="313a9b9c-4454-4927-a218-761e791cf156" path="/var/lib/kubelet/pods/313a9b9c-4454-4927-a218-761e791cf156/volumes" Dec 05 11:07:30.673667 master-0 kubenswrapper[24928]: I1205 11:07:30.668320 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-ec941-default-external-api-0" podStartSLOduration=7.6682913169999996 podStartE2EDuration="7.668291317s" podCreationTimestamp="2025-12-05 11:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:07:30.624203091 +0000 UTC m=+1210.627396942" watchObservedRunningTime="2025-12-05 11:07:30.668291317 +0000 UTC m=+1210.671485178" Dec 05 11:07:30.700509 master-0 kubenswrapper[24928]: I1205 11:07:30.695924 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cf777b7bc-z7q5s"] Dec 05 11:07:30.700509 master-0 kubenswrapper[24928]: I1205 11:07:30.696237 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" podUID="635e41cc-7a74-43a1-b4c8-2b069496dc3a" containerName="dnsmasq-dns" containerID="cri-o://99bfdf0f903af317086040622d17c3c36a8d8bb525ea8a062a787c708b7db1c7" gracePeriod=10 Dec 05 11:07:30.780704 master-0 kubenswrapper[24928]: I1205 11:07:30.780643 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Dec 05 11:07:31.071598 master-0 kubenswrapper[24928]: I1205 11:07:31.070169 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"1458275a-4e99-4257-88e5-e2ef36469ffc","Type":"ContainerStarted","Data":"6ade7ffee85688648a4e43accd8bd67971c62ade59469b78d5282455a37057d5"} Dec 05 11:07:31.075657 master-0 kubenswrapper[24928]: I1205 11:07:31.075258 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"a44ef9f2-4b09-4875-b3d3-4ff3508980a1","Type":"ContainerStarted","Data":"f7fffac54b775f629656bc7521caecec1761dee195cb2eb6a8f4d496f8a00daf"} Dec 05 11:07:31.089839 master-0 kubenswrapper[24928]: I1205 11:07:31.089771 24928 generic.go:334] "Generic (PLEG): container finished" podID="635e41cc-7a74-43a1-b4c8-2b069496dc3a" containerID="99bfdf0f903af317086040622d17c3c36a8d8bb525ea8a062a787c708b7db1c7" exitCode=0 Dec 05 11:07:31.090084 master-0 kubenswrapper[24928]: I1205 11:07:31.089938 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" event={"ID":"635e41cc-7a74-43a1-b4c8-2b069496dc3a","Type":"ContainerDied","Data":"99bfdf0f903af317086040622d17c3c36a8d8bb525ea8a062a787c708b7db1c7"} Dec 05 11:07:31.476886 master-0 kubenswrapper[24928]: I1205 11:07:31.476838 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:07:31.681409 master-0 kubenswrapper[24928]: I1205 11:07:31.681318 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-sb\") pod \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " Dec 05 11:07:31.681950 master-0 kubenswrapper[24928]: I1205 11:07:31.681542 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-svc\") pod \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " Dec 05 11:07:31.681950 master-0 kubenswrapper[24928]: I1205 11:07:31.681660 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pzjc\" (UniqueName: \"kubernetes.io/projected/635e41cc-7a74-43a1-b4c8-2b069496dc3a-kube-api-access-2pzjc\") pod \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " Dec 05 11:07:31.681950 master-0 kubenswrapper[24928]: I1205 11:07:31.681702 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-config\") pod \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " Dec 05 11:07:31.681950 master-0 kubenswrapper[24928]: I1205 11:07:31.681760 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-nb\") pod \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " Dec 05 11:07:31.681950 master-0 kubenswrapper[24928]: I1205 11:07:31.681779 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-swift-storage-0\") pod \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\" (UID: \"635e41cc-7a74-43a1-b4c8-2b069496dc3a\") " Dec 05 11:07:31.693656 master-0 kubenswrapper[24928]: I1205 11:07:31.693369 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/635e41cc-7a74-43a1-b4c8-2b069496dc3a-kube-api-access-2pzjc" (OuterVolumeSpecName: "kube-api-access-2pzjc") pod "635e41cc-7a74-43a1-b4c8-2b069496dc3a" (UID: "635e41cc-7a74-43a1-b4c8-2b069496dc3a"). InnerVolumeSpecName "kube-api-access-2pzjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:07:31.763453 master-0 kubenswrapper[24928]: I1205 11:07:31.757935 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "635e41cc-7a74-43a1-b4c8-2b069496dc3a" (UID: "635e41cc-7a74-43a1-b4c8-2b069496dc3a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:07:31.763453 master-0 kubenswrapper[24928]: I1205 11:07:31.759380 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-config" (OuterVolumeSpecName: "config") pod "635e41cc-7a74-43a1-b4c8-2b069496dc3a" (UID: "635e41cc-7a74-43a1-b4c8-2b069496dc3a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:07:31.763453 master-0 kubenswrapper[24928]: I1205 11:07:31.763386 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "635e41cc-7a74-43a1-b4c8-2b069496dc3a" (UID: "635e41cc-7a74-43a1-b4c8-2b069496dc3a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:07:31.773501 master-0 kubenswrapper[24928]: I1205 11:07:31.772972 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "635e41cc-7a74-43a1-b4c8-2b069496dc3a" (UID: "635e41cc-7a74-43a1-b4c8-2b069496dc3a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:07:31.784730 master-0 kubenswrapper[24928]: I1205 11:07:31.784629 24928 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-swift-storage-0\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:31.784730 master-0 kubenswrapper[24928]: I1205 11:07:31.784675 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-sb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:31.784730 master-0 kubenswrapper[24928]: I1205 11:07:31.784687 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:31.784730 master-0 kubenswrapper[24928]: I1205 11:07:31.784697 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pzjc\" (UniqueName: \"kubernetes.io/projected/635e41cc-7a74-43a1-b4c8-2b069496dc3a-kube-api-access-2pzjc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:31.784730 master-0 kubenswrapper[24928]: I1205 11:07:31.784708 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:31.802944 master-0 kubenswrapper[24928]: I1205 11:07:31.802859 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "635e41cc-7a74-43a1-b4c8-2b069496dc3a" (UID: "635e41cc-7a74-43a1-b4c8-2b069496dc3a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:07:31.887358 master-0 kubenswrapper[24928]: I1205 11:07:31.887274 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/635e41cc-7a74-43a1-b4c8-2b069496dc3a-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:07:32.109959 master-0 kubenswrapper[24928]: I1205 11:07:32.109881 24928 generic.go:334] "Generic (PLEG): container finished" podID="a44ef9f2-4b09-4875-b3d3-4ff3508980a1" containerID="1204be8b1fc6940ec90264ab60ffd822e7d4b445a14a0ffc705b3ae70eaf15c4" exitCode=0 Dec 05 11:07:32.110208 master-0 kubenswrapper[24928]: I1205 11:07:32.109962 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"a44ef9f2-4b09-4875-b3d3-4ff3508980a1","Type":"ContainerDied","Data":"1204be8b1fc6940ec90264ab60ffd822e7d4b445a14a0ffc705b3ae70eaf15c4"} Dec 05 11:07:32.112667 master-0 kubenswrapper[24928]: I1205 11:07:32.112592 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" event={"ID":"635e41cc-7a74-43a1-b4c8-2b069496dc3a","Type":"ContainerDied","Data":"094ae3dc207da40dce773f9a423aef73ae6d39279ba3830995ec7f15c8805c38"} Dec 05 11:07:32.112667 master-0 kubenswrapper[24928]: I1205 11:07:32.112644 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cf777b7bc-z7q5s" Dec 05 11:07:32.112667 master-0 kubenswrapper[24928]: I1205 11:07:32.112675 24928 scope.go:117] "RemoveContainer" containerID="99bfdf0f903af317086040622d17c3c36a8d8bb525ea8a062a787c708b7db1c7" Dec 05 11:07:32.115530 master-0 kubenswrapper[24928]: I1205 11:07:32.115468 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-ec941-default-internal-api-0" event={"ID":"1458275a-4e99-4257-88e5-e2ef36469ffc","Type":"ContainerStarted","Data":"a94ed81ed23ccccf1026c53a2c89de517bf6909412ebe8d24e1ce9a4eaf493a7"} Dec 05 11:07:32.142407 master-0 kubenswrapper[24928]: I1205 11:07:32.142332 24928 scope.go:117] "RemoveContainer" containerID="cca05638df01834d73cf5f18ee70814cf850b91358bd23f1a178f51e9764d727" Dec 05 11:07:32.199203 master-0 kubenswrapper[24928]: I1205 11:07:32.199127 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cf777b7bc-z7q5s"] Dec 05 11:07:32.232823 master-0 kubenswrapper[24928]: I1205 11:07:32.232755 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cf777b7bc-z7q5s"] Dec 05 11:07:33.129661 master-0 kubenswrapper[24928]: I1205 11:07:33.129588 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"a44ef9f2-4b09-4875-b3d3-4ff3508980a1","Type":"ContainerStarted","Data":"63c443590b18eba3002ef0251fc5c21932e8f644c9ae566c3d20fec17b770778"} Dec 05 11:07:33.425866 master-0 kubenswrapper[24928]: I1205 11:07:33.424880 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-ec941-default-internal-api-0" podStartSLOduration=9.424856069 podStartE2EDuration="9.424856069s" podCreationTimestamp="2025-12-05 11:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:07:32.221572431 +0000 UTC m=+1212.224766282" watchObservedRunningTime="2025-12-05 11:07:33.424856069 +0000 UTC m=+1213.428049940" Dec 05 11:07:34.157480 master-0 kubenswrapper[24928]: I1205 11:07:34.154685 24928 generic.go:334] "Generic (PLEG): container finished" podID="a44ef9f2-4b09-4875-b3d3-4ff3508980a1" containerID="63c443590b18eba3002ef0251fc5c21932e8f644c9ae566c3d20fec17b770778" exitCode=0 Dec 05 11:07:34.157480 master-0 kubenswrapper[24928]: I1205 11:07:34.154747 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"a44ef9f2-4b09-4875-b3d3-4ff3508980a1","Type":"ContainerDied","Data":"63c443590b18eba3002ef0251fc5c21932e8f644c9ae566c3d20fec17b770778"} Dec 05 11:07:34.221270 master-0 kubenswrapper[24928]: I1205 11:07:34.221206 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="635e41cc-7a74-43a1-b4c8-2b069496dc3a" path="/var/lib/kubelet/pods/635e41cc-7a74-43a1-b4c8-2b069496dc3a/volumes" Dec 05 11:07:35.234878 master-0 kubenswrapper[24928]: I1205 11:07:35.233179 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:35.234878 master-0 kubenswrapper[24928]: I1205 11:07:35.233265 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:35.284737 master-0 kubenswrapper[24928]: I1205 11:07:35.284036 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:35.288637 master-0 kubenswrapper[24928]: I1205 11:07:35.288581 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:36.188036 master-0 kubenswrapper[24928]: I1205 11:07:36.187900 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:36.188391 master-0 kubenswrapper[24928]: I1205 11:07:36.188053 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:38.264479 master-0 kubenswrapper[24928]: I1205 11:07:38.264408 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:38.265296 master-0 kubenswrapper[24928]: I1205 11:07:38.265275 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:07:38.271497 master-0 kubenswrapper[24928]: I1205 11:07:38.271438 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-ec941-default-external-api-0" Dec 05 11:07:38.607813 master-0 kubenswrapper[24928]: I1205 11:07:38.607682 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:38.609747 master-0 kubenswrapper[24928]: I1205 11:07:38.609667 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:38.645840 master-0 kubenswrapper[24928]: I1205 11:07:38.645735 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:38.659128 master-0 kubenswrapper[24928]: I1205 11:07:38.659032 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:39.227971 master-0 kubenswrapper[24928]: I1205 11:07:39.227899 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:39.227971 master-0 kubenswrapper[24928]: I1205 11:07:39.227949 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:41.256172 master-0 kubenswrapper[24928]: I1205 11:07:41.256099 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:07:41.256172 master-0 kubenswrapper[24928]: I1205 11:07:41.256152 24928 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 11:07:41.301260 master-0 kubenswrapper[24928]: I1205 11:07:41.301177 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:41.301533 master-0 kubenswrapper[24928]: I1205 11:07:41.301311 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-ec941-default-internal-api-0" Dec 05 11:07:46.331739 master-0 kubenswrapper[24928]: I1205 11:07:46.331644 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"a44ef9f2-4b09-4875-b3d3-4ff3508980a1","Type":"ContainerStarted","Data":"6c109d1f1e066fe5162c27a49c5e3ab8fb71ad4bc02a62a1ebefe72172cec26e"} Dec 05 11:07:47.350048 master-0 kubenswrapper[24928]: I1205 11:07:47.349971 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fjthc" event={"ID":"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578","Type":"ContainerStarted","Data":"5a5890da362f9de00fc5e6f60e48e89a3a543b86628b0d169ed86c857200fe1d"} Dec 05 11:07:47.367092 master-0 kubenswrapper[24928]: I1205 11:07:47.355824 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"a44ef9f2-4b09-4875-b3d3-4ff3508980a1","Type":"ContainerStarted","Data":"6fda94ceed7856076d26062035819af981376c67ebdeced184f5d80361c00b21"} Dec 05 11:07:47.367092 master-0 kubenswrapper[24928]: I1205 11:07:47.356161 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"a44ef9f2-4b09-4875-b3d3-4ff3508980a1","Type":"ContainerStarted","Data":"9d33b57df88ebaed2ee5f118f549ccb87927badcd473b80780d6931491670e1e"} Dec 05 11:07:47.378049 master-0 kubenswrapper[24928]: I1205 11:07:47.377952 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-fjthc" podStartSLOduration=2.503956269 podStartE2EDuration="19.377937518s" podCreationTimestamp="2025-12-05 11:07:28 +0000 UTC" firstStartedPulling="2025-12-05 11:07:29.172498206 +0000 UTC m=+1209.175692057" lastFinishedPulling="2025-12-05 11:07:46.046479455 +0000 UTC m=+1226.049673306" observedRunningTime="2025-12-05 11:07:47.374868064 +0000 UTC m=+1227.378061935" watchObservedRunningTime="2025-12-05 11:07:47.377937518 +0000 UTC m=+1227.381131359" Dec 05 11:07:48.416581 master-0 kubenswrapper[24928]: I1205 11:07:48.416500 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"a44ef9f2-4b09-4875-b3d3-4ff3508980a1","Type":"ContainerStarted","Data":"6575a010d134f81f55c6487aa6ce3f08f0c393ae0528aac922134af100c745ae"} Dec 05 11:07:48.416581 master-0 kubenswrapper[24928]: I1205 11:07:48.416587 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"a44ef9f2-4b09-4875-b3d3-4ff3508980a1","Type":"ContainerStarted","Data":"1cd0681924c7d8fadd558beb7117ad749f89978005c0682d0eb48213ec34e427"} Dec 05 11:07:48.417392 master-0 kubenswrapper[24928]: I1205 11:07:48.416789 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Dec 05 11:07:48.609454 master-0 kubenswrapper[24928]: I1205 11:07:48.607233 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-0" podStartSLOduration=19.60719554 podStartE2EDuration="19.60719554s" podCreationTimestamp="2025-12-05 11:07:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:07:48.577788742 +0000 UTC m=+1228.580982603" watchObservedRunningTime="2025-12-05 11:07:48.60719554 +0000 UTC m=+1228.610389391" Dec 05 11:07:49.430124 master-0 kubenswrapper[24928]: I1205 11:07:49.430058 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Dec 05 11:07:49.645336 master-0 kubenswrapper[24928]: I1205 11:07:49.645204 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Dec 05 11:07:49.646103 master-0 kubenswrapper[24928]: I1205 11:07:49.646060 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Dec 05 11:07:49.646195 master-0 kubenswrapper[24928]: I1205 11:07:49.646107 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Dec 05 11:07:49.646195 master-0 kubenswrapper[24928]: I1205 11:07:49.646123 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Dec 05 11:07:49.676020 master-0 kubenswrapper[24928]: I1205 11:07:49.675957 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ironic-inspector-0" Dec 05 11:07:49.681096 master-0 kubenswrapper[24928]: I1205 11:07:49.680950 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ironic-inspector-0" Dec 05 11:07:50.458354 master-0 kubenswrapper[24928]: I1205 11:07:50.458253 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Dec 05 11:07:50.463472 master-0 kubenswrapper[24928]: I1205 11:07:50.463387 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Dec 05 11:07:50.490807 master-0 kubenswrapper[24928]: I1205 11:07:50.482686 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Dec 05 11:07:51.455651 master-0 kubenswrapper[24928]: I1205 11:07:51.455594 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Dec 05 11:08:02.639374 master-0 kubenswrapper[24928]: I1205 11:08:02.639300 24928 generic.go:334] "Generic (PLEG): container finished" podID="7b68cc45-2d4a-4377-ba5c-e35ecd4ed578" containerID="5a5890da362f9de00fc5e6f60e48e89a3a543b86628b0d169ed86c857200fe1d" exitCode=0 Dec 05 11:08:02.639374 master-0 kubenswrapper[24928]: I1205 11:08:02.639358 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fjthc" event={"ID":"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578","Type":"ContainerDied","Data":"5a5890da362f9de00fc5e6f60e48e89a3a543b86628b0d169ed86c857200fe1d"} Dec 05 11:08:04.114887 master-0 kubenswrapper[24928]: I1205 11:08:04.114822 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:08:04.241923 master-0 kubenswrapper[24928]: I1205 11:08:04.241829 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4clb\" (UniqueName: \"kubernetes.io/projected/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-kube-api-access-t4clb\") pod \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " Dec 05 11:08:04.242209 master-0 kubenswrapper[24928]: I1205 11:08:04.242116 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-scripts\") pod \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " Dec 05 11:08:04.242209 master-0 kubenswrapper[24928]: I1205 11:08:04.242179 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-combined-ca-bundle\") pod \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " Dec 05 11:08:04.242303 master-0 kubenswrapper[24928]: I1205 11:08:04.242225 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-config-data\") pod \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\" (UID: \"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578\") " Dec 05 11:08:04.253954 master-0 kubenswrapper[24928]: I1205 11:08:04.253850 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-scripts" (OuterVolumeSpecName: "scripts") pod "7b68cc45-2d4a-4377-ba5c-e35ecd4ed578" (UID: "7b68cc45-2d4a-4377-ba5c-e35ecd4ed578"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:04.254215 master-0 kubenswrapper[24928]: I1205 11:08:04.254075 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-kube-api-access-t4clb" (OuterVolumeSpecName: "kube-api-access-t4clb") pod "7b68cc45-2d4a-4377-ba5c-e35ecd4ed578" (UID: "7b68cc45-2d4a-4377-ba5c-e35ecd4ed578"). InnerVolumeSpecName "kube-api-access-t4clb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:04.274554 master-0 kubenswrapper[24928]: I1205 11:08:04.274397 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-config-data" (OuterVolumeSpecName: "config-data") pod "7b68cc45-2d4a-4377-ba5c-e35ecd4ed578" (UID: "7b68cc45-2d4a-4377-ba5c-e35ecd4ed578"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:04.283941 master-0 kubenswrapper[24928]: I1205 11:08:04.283874 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b68cc45-2d4a-4377-ba5c-e35ecd4ed578" (UID: "7b68cc45-2d4a-4377-ba5c-e35ecd4ed578"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:04.345401 master-0 kubenswrapper[24928]: I1205 11:08:04.345315 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4clb\" (UniqueName: \"kubernetes.io/projected/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-kube-api-access-t4clb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:04.345401 master-0 kubenswrapper[24928]: I1205 11:08:04.345379 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:04.345401 master-0 kubenswrapper[24928]: I1205 11:08:04.345392 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:04.345401 master-0 kubenswrapper[24928]: I1205 11:08:04.345402 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:04.680633 master-0 kubenswrapper[24928]: I1205 11:08:04.680557 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-fjthc" event={"ID":"7b68cc45-2d4a-4377-ba5c-e35ecd4ed578","Type":"ContainerDied","Data":"e7eb47c961ccc3f646a574f2974202d7027782a909645c7173030e1d846662b1"} Dec 05 11:08:04.680633 master-0 kubenswrapper[24928]: I1205 11:08:04.680624 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7eb47c961ccc3f646a574f2974202d7027782a909645c7173030e1d846662b1" Dec 05 11:08:04.680923 master-0 kubenswrapper[24928]: I1205 11:08:04.680697 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-fjthc" Dec 05 11:08:04.879720 master-0 kubenswrapper[24928]: I1205 11:08:04.879653 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 11:08:04.880194 master-0 kubenswrapper[24928]: E1205 11:08:04.880159 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="635e41cc-7a74-43a1-b4c8-2b069496dc3a" containerName="init" Dec 05 11:08:04.880194 master-0 kubenswrapper[24928]: I1205 11:08:04.880189 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="635e41cc-7a74-43a1-b4c8-2b069496dc3a" containerName="init" Dec 05 11:08:04.880276 master-0 kubenswrapper[24928]: E1205 11:08:04.880205 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b68cc45-2d4a-4377-ba5c-e35ecd4ed578" containerName="nova-cell0-conductor-db-sync" Dec 05 11:08:04.880276 master-0 kubenswrapper[24928]: I1205 11:08:04.880213 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b68cc45-2d4a-4377-ba5c-e35ecd4ed578" containerName="nova-cell0-conductor-db-sync" Dec 05 11:08:04.880276 master-0 kubenswrapper[24928]: E1205 11:08:04.880253 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="635e41cc-7a74-43a1-b4c8-2b069496dc3a" containerName="dnsmasq-dns" Dec 05 11:08:04.880276 master-0 kubenswrapper[24928]: I1205 11:08:04.880263 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="635e41cc-7a74-43a1-b4c8-2b069496dc3a" containerName="dnsmasq-dns" Dec 05 11:08:04.880539 master-0 kubenswrapper[24928]: I1205 11:08:04.880513 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b68cc45-2d4a-4377-ba5c-e35ecd4ed578" containerName="nova-cell0-conductor-db-sync" Dec 05 11:08:04.880589 master-0 kubenswrapper[24928]: I1205 11:08:04.880549 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="635e41cc-7a74-43a1-b4c8-2b069496dc3a" containerName="dnsmasq-dns" Dec 05 11:08:04.881334 master-0 kubenswrapper[24928]: I1205 11:08:04.881302 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:04.883892 master-0 kubenswrapper[24928]: I1205 11:08:04.883854 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 11:08:04.922912 master-0 kubenswrapper[24928]: I1205 11:08:04.922852 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 11:08:04.958642 master-0 kubenswrapper[24928]: I1205 11:08:04.958414 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww59b\" (UniqueName: \"kubernetes.io/projected/2c07350c-afb9-49aa-b149-cf6d80491725-kube-api-access-ww59b\") pod \"nova-cell0-conductor-0\" (UID: \"2c07350c-afb9-49aa-b149-cf6d80491725\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:04.958890 master-0 kubenswrapper[24928]: I1205 11:08:04.958689 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c07350c-afb9-49aa-b149-cf6d80491725-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2c07350c-afb9-49aa-b149-cf6d80491725\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:04.958890 master-0 kubenswrapper[24928]: I1205 11:08:04.958879 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c07350c-afb9-49aa-b149-cf6d80491725-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2c07350c-afb9-49aa-b149-cf6d80491725\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:05.061237 master-0 kubenswrapper[24928]: I1205 11:08:05.061161 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww59b\" (UniqueName: \"kubernetes.io/projected/2c07350c-afb9-49aa-b149-cf6d80491725-kube-api-access-ww59b\") pod \"nova-cell0-conductor-0\" (UID: \"2c07350c-afb9-49aa-b149-cf6d80491725\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:05.061492 master-0 kubenswrapper[24928]: I1205 11:08:05.061299 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c07350c-afb9-49aa-b149-cf6d80491725-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2c07350c-afb9-49aa-b149-cf6d80491725\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:05.061492 master-0 kubenswrapper[24928]: I1205 11:08:05.061473 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c07350c-afb9-49aa-b149-cf6d80491725-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2c07350c-afb9-49aa-b149-cf6d80491725\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:05.065725 master-0 kubenswrapper[24928]: I1205 11:08:05.065439 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c07350c-afb9-49aa-b149-cf6d80491725-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"2c07350c-afb9-49aa-b149-cf6d80491725\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:05.065725 master-0 kubenswrapper[24928]: I1205 11:08:05.065571 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c07350c-afb9-49aa-b149-cf6d80491725-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"2c07350c-afb9-49aa-b149-cf6d80491725\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:05.083074 master-0 kubenswrapper[24928]: I1205 11:08:05.083006 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww59b\" (UniqueName: \"kubernetes.io/projected/2c07350c-afb9-49aa-b149-cf6d80491725-kube-api-access-ww59b\") pod \"nova-cell0-conductor-0\" (UID: \"2c07350c-afb9-49aa-b149-cf6d80491725\") " pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:05.216957 master-0 kubenswrapper[24928]: I1205 11:08:05.216832 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:05.727564 master-0 kubenswrapper[24928]: W1205 11:08:05.727491 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c07350c_afb9_49aa_b149_cf6d80491725.slice/crio-742615f62a60795362b82d16e9870960004ce4197089bc05366b4bb7f6e855d8 WatchSource:0}: Error finding container 742615f62a60795362b82d16e9870960004ce4197089bc05366b4bb7f6e855d8: Status 404 returned error can't find the container with id 742615f62a60795362b82d16e9870960004ce4197089bc05366b4bb7f6e855d8 Dec 05 11:08:05.735846 master-0 kubenswrapper[24928]: I1205 11:08:05.735635 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 11:08:06.705591 master-0 kubenswrapper[24928]: I1205 11:08:06.705506 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2c07350c-afb9-49aa-b149-cf6d80491725","Type":"ContainerStarted","Data":"e1cd1a5c46f22a59bf7d9586394b7fe130f8eb2a805de1cf4124692f632744e1"} Dec 05 11:08:06.705591 master-0 kubenswrapper[24928]: I1205 11:08:06.705574 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"2c07350c-afb9-49aa-b149-cf6d80491725","Type":"ContainerStarted","Data":"742615f62a60795362b82d16e9870960004ce4197089bc05366b4bb7f6e855d8"} Dec 05 11:08:06.707454 master-0 kubenswrapper[24928]: I1205 11:08:06.707228 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:06.735204 master-0 kubenswrapper[24928]: I1205 11:08:06.735088 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.7350682600000003 podStartE2EDuration="2.73506826s" podCreationTimestamp="2025-12-05 11:08:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:06.727290301 +0000 UTC m=+1246.730484162" watchObservedRunningTime="2025-12-05 11:08:06.73506826 +0000 UTC m=+1246.738262111" Dec 05 11:08:10.258150 master-0 kubenswrapper[24928]: I1205 11:08:10.258083 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 11:08:11.028337 master-0 kubenswrapper[24928]: I1205 11:08:11.028256 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-8tsgs"] Dec 05 11:08:11.030434 master-0 kubenswrapper[24928]: I1205 11:08:11.030366 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.033697 master-0 kubenswrapper[24928]: I1205 11:08:11.033649 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 05 11:08:11.034402 master-0 kubenswrapper[24928]: I1205 11:08:11.034373 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 05 11:08:11.043378 master-0 kubenswrapper[24928]: I1205 11:08:11.043315 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-8tsgs"] Dec 05 11:08:11.110955 master-0 kubenswrapper[24928]: I1205 11:08:11.110632 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-scripts\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.110955 master-0 kubenswrapper[24928]: I1205 11:08:11.110777 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-config-data\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.110955 master-0 kubenswrapper[24928]: I1205 11:08:11.110865 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.111251 master-0 kubenswrapper[24928]: I1205 11:08:11.110992 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj8kf\" (UniqueName: \"kubernetes.io/projected/f6dde2d1-2c98-48c1-ac60-02129e7543b3-kube-api-access-hj8kf\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.184624 master-0 kubenswrapper[24928]: I1205 11:08:11.184548 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-compute-ironic-compute-0"] Dec 05 11:08:11.186506 master-0 kubenswrapper[24928]: I1205 11:08:11.186467 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:11.189177 master-0 kubenswrapper[24928]: I1205 11:08:11.189130 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-ironic-compute-config-data" Dec 05 11:08:11.218356 master-0 kubenswrapper[24928]: I1205 11:08:11.218301 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.218613 master-0 kubenswrapper[24928]: I1205 11:08:11.218555 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj8kf\" (UniqueName: \"kubernetes.io/projected/f6dde2d1-2c98-48c1-ac60-02129e7543b3-kube-api-access-hj8kf\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.218678 master-0 kubenswrapper[24928]: I1205 11:08:11.218631 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-scripts\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.218818 master-0 kubenswrapper[24928]: I1205 11:08:11.218777 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-config-data\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.227537 master-0 kubenswrapper[24928]: I1205 11:08:11.224639 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-config-data\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.231732 master-0 kubenswrapper[24928]: I1205 11:08:11.227731 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-compute-ironic-compute-0"] Dec 05 11:08:11.235523 master-0 kubenswrapper[24928]: I1205 11:08:11.232238 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-scripts\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.240730 master-0 kubenswrapper[24928]: I1205 11:08:11.240688 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.355987 master-0 kubenswrapper[24928]: I1205 11:08:11.355880 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj8kf\" (UniqueName: \"kubernetes.io/projected/f6dde2d1-2c98-48c1-ac60-02129e7543b3-kube-api-access-hj8kf\") pod \"nova-cell0-cell-mapping-8tsgs\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.357694 master-0 kubenswrapper[24928]: I1205 11:08:11.357303 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:11.424489 master-0 kubenswrapper[24928]: I1205 11:08:11.424367 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md4wt\" (UniqueName: \"kubernetes.io/projected/fed98f34-47f9-4239-ac14-637b8ca68b73-kube-api-access-md4wt\") pod \"nova-cell1-compute-ironic-compute-0\" (UID: \"fed98f34-47f9-4239-ac14-637b8ca68b73\") " pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:11.424578 master-0 kubenswrapper[24928]: I1205 11:08:11.424537 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed98f34-47f9-4239-ac14-637b8ca68b73-config-data\") pod \"nova-cell1-compute-ironic-compute-0\" (UID: \"fed98f34-47f9-4239-ac14-637b8ca68b73\") " pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:11.424837 master-0 kubenswrapper[24928]: I1205 11:08:11.424788 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed98f34-47f9-4239-ac14-637b8ca68b73-combined-ca-bundle\") pod \"nova-cell1-compute-ironic-compute-0\" (UID: \"fed98f34-47f9-4239-ac14-637b8ca68b73\") " pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:11.529358 master-0 kubenswrapper[24928]: I1205 11:08:11.527085 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed98f34-47f9-4239-ac14-637b8ca68b73-combined-ca-bundle\") pod \"nova-cell1-compute-ironic-compute-0\" (UID: \"fed98f34-47f9-4239-ac14-637b8ca68b73\") " pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:11.529358 master-0 kubenswrapper[24928]: I1205 11:08:11.527259 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md4wt\" (UniqueName: \"kubernetes.io/projected/fed98f34-47f9-4239-ac14-637b8ca68b73-kube-api-access-md4wt\") pod \"nova-cell1-compute-ironic-compute-0\" (UID: \"fed98f34-47f9-4239-ac14-637b8ca68b73\") " pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:11.529358 master-0 kubenswrapper[24928]: I1205 11:08:11.527317 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed98f34-47f9-4239-ac14-637b8ca68b73-config-data\") pod \"nova-cell1-compute-ironic-compute-0\" (UID: \"fed98f34-47f9-4239-ac14-637b8ca68b73\") " pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:11.532758 master-0 kubenswrapper[24928]: I1205 11:08:11.530857 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fed98f34-47f9-4239-ac14-637b8ca68b73-config-data\") pod \"nova-cell1-compute-ironic-compute-0\" (UID: \"fed98f34-47f9-4239-ac14-637b8ca68b73\") " pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:11.552731 master-0 kubenswrapper[24928]: I1205 11:08:11.552344 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fed98f34-47f9-4239-ac14-637b8ca68b73-combined-ca-bundle\") pod \"nova-cell1-compute-ironic-compute-0\" (UID: \"fed98f34-47f9-4239-ac14-637b8ca68b73\") " pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:11.597386 master-0 kubenswrapper[24928]: I1205 11:08:11.596371 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md4wt\" (UniqueName: \"kubernetes.io/projected/fed98f34-47f9-4239-ac14-637b8ca68b73-kube-api-access-md4wt\") pod \"nova-cell1-compute-ironic-compute-0\" (UID: \"fed98f34-47f9-4239-ac14-637b8ca68b73\") " pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:11.621532 master-0 kubenswrapper[24928]: I1205 11:08:11.620468 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:11.631558 master-0 kubenswrapper[24928]: I1205 11:08:11.631329 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:08:11.636643 master-0 kubenswrapper[24928]: I1205 11:08:11.634273 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 11:08:11.642322 master-0 kubenswrapper[24928]: I1205 11:08:11.642168 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:08:11.644365 master-0 kubenswrapper[24928]: I1205 11:08:11.644310 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:11.646601 master-0 kubenswrapper[24928]: I1205 11:08:11.646562 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 11:08:11.674866 master-0 kubenswrapper[24928]: I1205 11:08:11.660555 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:08:11.674866 master-0 kubenswrapper[24928]: I1205 11:08:11.662413 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:08:11.674866 master-0 kubenswrapper[24928]: I1205 11:08:11.664892 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 11:08:11.674866 master-0 kubenswrapper[24928]: I1205 11:08:11.671197 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:11.682914 master-0 kubenswrapper[24928]: I1205 11:08:11.682804 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:08:11.695608 master-0 kubenswrapper[24928]: I1205 11:08:11.695154 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:08:11.707912 master-0 kubenswrapper[24928]: I1205 11:08:11.707829 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:11.734109 master-0 kubenswrapper[24928]: I1205 11:08:11.734047 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62fbcd8d-b895-4135-bc9a-339f4369e359-logs\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:11.734320 master-0 kubenswrapper[24928]: I1205 11:08:11.734212 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6fvg\" (UniqueName: \"kubernetes.io/projected/923a2187-d634-44ba-bb89-bd632b6e2908-kube-api-access-w6fvg\") pod \"nova-scheduler-0\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:11.734320 master-0 kubenswrapper[24928]: I1205 11:08:11.734274 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:11.734448 master-0 kubenswrapper[24928]: I1205 11:08:11.734347 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-config-data\") pod \"nova-scheduler-0\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:11.734448 master-0 kubenswrapper[24928]: I1205 11:08:11.734381 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-config-data\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:11.734448 master-0 kubenswrapper[24928]: I1205 11:08:11.734406 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nhx7\" (UniqueName: \"kubernetes.io/projected/62fbcd8d-b895-4135-bc9a-339f4369e359-kube-api-access-4nhx7\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:11.734567 master-0 kubenswrapper[24928]: I1205 11:08:11.734487 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:11.734567 master-0 kubenswrapper[24928]: I1205 11:08:11.734520 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vf7t\" (UniqueName: \"kubernetes.io/projected/3391ea32-ded0-4fc0-92b6-823023c6b9ca-kube-api-access-4vf7t\") pod \"nova-cell1-novncproxy-0\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:11.734637 master-0 kubenswrapper[24928]: I1205 11:08:11.734603 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:11.734673 master-0 kubenswrapper[24928]: I1205 11:08:11.734651 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:11.837140 master-0 kubenswrapper[24928]: I1205 11:08:11.837056 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-config-data\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:11.837140 master-0 kubenswrapper[24928]: I1205 11:08:11.837123 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nhx7\" (UniqueName: \"kubernetes.io/projected/62fbcd8d-b895-4135-bc9a-339f4369e359-kube-api-access-4nhx7\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:11.837340 master-0 kubenswrapper[24928]: I1205 11:08:11.837171 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:11.837340 master-0 kubenswrapper[24928]: I1205 11:08:11.837200 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vf7t\" (UniqueName: \"kubernetes.io/projected/3391ea32-ded0-4fc0-92b6-823023c6b9ca-kube-api-access-4vf7t\") pod \"nova-cell1-novncproxy-0\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:11.837340 master-0 kubenswrapper[24928]: I1205 11:08:11.837268 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:11.837340 master-0 kubenswrapper[24928]: I1205 11:08:11.837304 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:11.837340 master-0 kubenswrapper[24928]: I1205 11:08:11.837341 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62fbcd8d-b895-4135-bc9a-339f4369e359-logs\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:11.837542 master-0 kubenswrapper[24928]: I1205 11:08:11.837369 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6fvg\" (UniqueName: \"kubernetes.io/projected/923a2187-d634-44ba-bb89-bd632b6e2908-kube-api-access-w6fvg\") pod \"nova-scheduler-0\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:11.837542 master-0 kubenswrapper[24928]: I1205 11:08:11.837395 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:11.839380 master-0 kubenswrapper[24928]: I1205 11:08:11.839343 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62fbcd8d-b895-4135-bc9a-339f4369e359-logs\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:11.839765 master-0 kubenswrapper[24928]: I1205 11:08:11.839732 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-config-data\") pod \"nova-scheduler-0\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:11.843021 master-0 kubenswrapper[24928]: I1205 11:08:11.842983 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-config-data\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:11.843390 master-0 kubenswrapper[24928]: I1205 11:08:11.843349 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:11.843831 master-0 kubenswrapper[24928]: I1205 11:08:11.843741 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:11.845449 master-0 kubenswrapper[24928]: I1205 11:08:11.845398 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:11.846535 master-0 kubenswrapper[24928]: I1205 11:08:11.846491 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-config-data\") pod \"nova-scheduler-0\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:11.847621 master-0 kubenswrapper[24928]: I1205 11:08:11.847575 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:11.999870 master-0 kubenswrapper[24928]: I1205 11:08:11.999804 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nhx7\" (UniqueName: \"kubernetes.io/projected/62fbcd8d-b895-4135-bc9a-339f4369e359-kube-api-access-4nhx7\") pod \"nova-api-0\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " pod="openstack/nova-api-0" Dec 05 11:08:12.004913 master-0 kubenswrapper[24928]: I1205 11:08:12.004862 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vf7t\" (UniqueName: \"kubernetes.io/projected/3391ea32-ded0-4fc0-92b6-823023c6b9ca-kube-api-access-4vf7t\") pod \"nova-cell1-novncproxy-0\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:12.014408 master-0 kubenswrapper[24928]: I1205 11:08:12.014343 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:08:12.015058 master-0 kubenswrapper[24928]: I1205 11:08:12.015010 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6fvg\" (UniqueName: \"kubernetes.io/projected/923a2187-d634-44ba-bb89-bd632b6e2908-kube-api-access-w6fvg\") pod \"nova-scheduler-0\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:12.017718 master-0 kubenswrapper[24928]: I1205 11:08:12.017048 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:12.025011 master-0 kubenswrapper[24928]: I1205 11:08:12.024929 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-8tsgs"] Dec 05 11:08:12.027977 master-0 kubenswrapper[24928]: I1205 11:08:12.027203 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:08:12.039846 master-0 kubenswrapper[24928]: W1205 11:08:12.039784 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6dde2d1_2c98_48c1_ac60_02129e7543b3.slice/crio-483d9bcba42e40d48f087889b4fb4d5b0dafc97bd1a17b36ac01345aec501910 WatchSource:0}: Error finding container 483d9bcba42e40d48f087889b4fb4d5b0dafc97bd1a17b36ac01345aec501910: Status 404 returned error can't find the container with id 483d9bcba42e40d48f087889b4fb4d5b0dafc97bd1a17b36ac01345aec501910 Dec 05 11:08:12.057248 master-0 kubenswrapper[24928]: I1205 11:08:12.055667 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:12.058297 master-0 kubenswrapper[24928]: I1205 11:08:12.058042 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:12.069788 master-0 kubenswrapper[24928]: I1205 11:08:12.069347 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 11:08:12.098577 master-0 kubenswrapper[24928]: I1205 11:08:12.098415 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:12.472054 master-0 kubenswrapper[24928]: I1205 11:08:12.471998 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-logs\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.472717 master-0 kubenswrapper[24928]: I1205 11:08:12.472685 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-config-data\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.472794 master-0 kubenswrapper[24928]: I1205 11:08:12.472763 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rrmg\" (UniqueName: \"kubernetes.io/projected/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-kube-api-access-4rrmg\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.472834 master-0 kubenswrapper[24928]: I1205 11:08:12.472823 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.492900 master-0 kubenswrapper[24928]: W1205 11:08:12.489550 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfed98f34_47f9_4239_ac14_637b8ca68b73.slice/crio-4a9ad95f0507d011c8d12c09c98f8977b6fb354ae4f4bc43ddbfec35abb7cf71 WatchSource:0}: Error finding container 4a9ad95f0507d011c8d12c09c98f8977b6fb354ae4f4bc43ddbfec35abb7cf71: Status 404 returned error can't find the container with id 4a9ad95f0507d011c8d12c09c98f8977b6fb354ae4f4bc43ddbfec35abb7cf71 Dec 05 11:08:12.526487 master-0 kubenswrapper[24928]: I1205 11:08:12.523154 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tqtr9"] Dec 05 11:08:12.546903 master-0 kubenswrapper[24928]: I1205 11:08:12.545496 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-compute-ironic-compute-0"] Dec 05 11:08:12.546903 master-0 kubenswrapper[24928]: I1205 11:08:12.545612 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.548350 master-0 kubenswrapper[24928]: I1205 11:08:12.547760 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 05 11:08:12.553069 master-0 kubenswrapper[24928]: I1205 11:08:12.552316 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 11:08:12.554574 master-0 kubenswrapper[24928]: I1205 11:08:12.554516 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tqtr9"] Dec 05 11:08:12.575805 master-0 kubenswrapper[24928]: I1205 11:08:12.575541 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-logs\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.575805 master-0 kubenswrapper[24928]: I1205 11:08:12.575606 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-config-data\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.575805 master-0 kubenswrapper[24928]: I1205 11:08:12.575632 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-config-data\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.575805 master-0 kubenswrapper[24928]: I1205 11:08:12.575662 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.575805 master-0 kubenswrapper[24928]: I1205 11:08:12.575688 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rrmg\" (UniqueName: \"kubernetes.io/projected/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-kube-api-access-4rrmg\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.575805 master-0 kubenswrapper[24928]: I1205 11:08:12.575726 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.575805 master-0 kubenswrapper[24928]: I1205 11:08:12.575794 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kd7f\" (UniqueName: \"kubernetes.io/projected/cdfae6e2-22a3-4218-b6e9-2558a61bcead-kube-api-access-8kd7f\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.576538 master-0 kubenswrapper[24928]: I1205 11:08:12.575859 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-scripts\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.576686 master-0 kubenswrapper[24928]: I1205 11:08:12.576640 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-logs\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.586649 master-0 kubenswrapper[24928]: I1205 11:08:12.586459 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-config-data\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.602180 master-0 kubenswrapper[24928]: I1205 11:08:12.602134 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rrmg\" (UniqueName: \"kubernetes.io/projected/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-kube-api-access-4rrmg\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.604192 master-0 kubenswrapper[24928]: I1205 11:08:12.603220 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " pod="openstack/nova-metadata-0" Dec 05 11:08:12.699644 master-0 kubenswrapper[24928]: I1205 11:08:12.699329 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76dc4d7685-49kgr"] Dec 05 11:08:12.703559 master-0 kubenswrapper[24928]: I1205 11:08:12.703492 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-config-data\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.703734 master-0 kubenswrapper[24928]: I1205 11:08:12.703623 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.708816 master-0 kubenswrapper[24928]: I1205 11:08:12.703903 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kd7f\" (UniqueName: \"kubernetes.io/projected/cdfae6e2-22a3-4218-b6e9-2558a61bcead-kube-api-access-8kd7f\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.708816 master-0 kubenswrapper[24928]: I1205 11:08:12.704085 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-scripts\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.724234 master-0 kubenswrapper[24928]: I1205 11:08:12.719232 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.724234 master-0 kubenswrapper[24928]: I1205 11:08:12.721097 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.743470 master-0 kubenswrapper[24928]: I1205 11:08:12.728335 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-scripts\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.743470 master-0 kubenswrapper[24928]: I1205 11:08:12.739411 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76dc4d7685-49kgr"] Dec 05 11:08:12.743470 master-0 kubenswrapper[24928]: I1205 11:08:12.741765 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-config-data\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:12.807976 master-0 kubenswrapper[24928]: I1205 11:08:12.806915 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-nb\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.807976 master-0 kubenswrapper[24928]: I1205 11:08:12.807013 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-config\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.807976 master-0 kubenswrapper[24928]: I1205 11:08:12.807329 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-svc\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.807976 master-0 kubenswrapper[24928]: I1205 11:08:12.807406 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmq2h\" (UniqueName: \"kubernetes.io/projected/9e744111-e3f8-4e33-8fc5-eb86882de858-kube-api-access-xmq2h\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.807976 master-0 kubenswrapper[24928]: I1205 11:08:12.807491 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-swift-storage-0\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.807976 master-0 kubenswrapper[24928]: I1205 11:08:12.807712 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-sb\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.815737 master-0 kubenswrapper[24928]: I1205 11:08:12.815655 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-compute-ironic-compute-0" event={"ID":"fed98f34-47f9-4239-ac14-637b8ca68b73","Type":"ContainerStarted","Data":"4a9ad95f0507d011c8d12c09c98f8977b6fb354ae4f4bc43ddbfec35abb7cf71"} Dec 05 11:08:12.824909 master-0 kubenswrapper[24928]: I1205 11:08:12.820000 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8tsgs" event={"ID":"f6dde2d1-2c98-48c1-ac60-02129e7543b3","Type":"ContainerStarted","Data":"483d9bcba42e40d48f087889b4fb4d5b0dafc97bd1a17b36ac01345aec501910"} Dec 05 11:08:12.824909 master-0 kubenswrapper[24928]: I1205 11:08:12.823269 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:12.910678 master-0 kubenswrapper[24928]: I1205 11:08:12.910626 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-svc\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.911190 master-0 kubenswrapper[24928]: I1205 11:08:12.911100 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmq2h\" (UniqueName: \"kubernetes.io/projected/9e744111-e3f8-4e33-8fc5-eb86882de858-kube-api-access-xmq2h\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.911371 master-0 kubenswrapper[24928]: I1205 11:08:12.911353 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-swift-storage-0\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.911561 master-0 kubenswrapper[24928]: I1205 11:08:12.911542 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-sb\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.912003 master-0 kubenswrapper[24928]: I1205 11:08:12.911985 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-nb\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.912330 master-0 kubenswrapper[24928]: I1205 11:08:12.912286 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-swift-storage-0\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.912880 master-0 kubenswrapper[24928]: I1205 11:08:12.912750 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-config\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.913028 master-0 kubenswrapper[24928]: I1205 11:08:12.912987 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-nb\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.913524 master-0 kubenswrapper[24928]: I1205 11:08:12.913486 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-config\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.913742 master-0 kubenswrapper[24928]: I1205 11:08:12.913703 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-svc\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:12.915187 master-0 kubenswrapper[24928]: I1205 11:08:12.914955 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-sb\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:13.233462 master-0 kubenswrapper[24928]: I1205 11:08:13.230654 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kd7f\" (UniqueName: \"kubernetes.io/projected/cdfae6e2-22a3-4218-b6e9-2558a61bcead-kube-api-access-8kd7f\") pod \"nova-cell1-conductor-db-sync-tqtr9\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:13.309100 master-0 kubenswrapper[24928]: I1205 11:08:13.309038 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmq2h\" (UniqueName: \"kubernetes.io/projected/9e744111-e3f8-4e33-8fc5-eb86882de858-kube-api-access-xmq2h\") pod \"dnsmasq-dns-76dc4d7685-49kgr\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:13.376461 master-0 kubenswrapper[24928]: I1205 11:08:13.375289 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:13.417359 master-0 kubenswrapper[24928]: I1205 11:08:13.417276 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:13.445130 master-0 kubenswrapper[24928]: I1205 11:08:13.443888 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:08:13.604812 master-0 kubenswrapper[24928]: W1205 11:08:13.604731 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod923a2187_d634_44ba_bb89_bd632b6e2908.slice/crio-1883ca908e07a5ad552278d8a5ef796feb8a5eac5cbe069929edbd04b5a61254 WatchSource:0}: Error finding container 1883ca908e07a5ad552278d8a5ef796feb8a5eac5cbe069929edbd04b5a61254: Status 404 returned error can't find the container with id 1883ca908e07a5ad552278d8a5ef796feb8a5eac5cbe069929edbd04b5a61254 Dec 05 11:08:13.619973 master-0 kubenswrapper[24928]: I1205 11:08:13.619916 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:08:13.642109 master-0 kubenswrapper[24928]: I1205 11:08:13.642037 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:13.656901 master-0 kubenswrapper[24928]: W1205 11:08:13.656782 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62fbcd8d_b895_4135_bc9a_339f4369e359.slice/crio-0bd75a6069efdd0216a63272838195c490886efb465319fb9739be80cbcd30d9 WatchSource:0}: Error finding container 0bd75a6069efdd0216a63272838195c490886efb465319fb9739be80cbcd30d9: Status 404 returned error can't find the container with id 0bd75a6069efdd0216a63272838195c490886efb465319fb9739be80cbcd30d9 Dec 05 11:08:13.829863 master-0 kubenswrapper[24928]: I1205 11:08:13.829791 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:13.857760 master-0 kubenswrapper[24928]: I1205 11:08:13.856518 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8tsgs" event={"ID":"f6dde2d1-2c98-48c1-ac60-02129e7543b3","Type":"ContainerStarted","Data":"fa1dbb07b11ef157ddc386720c998f0bd948fbc44b183a8e3569217c865c86c9"} Dec 05 11:08:13.883526 master-0 kubenswrapper[24928]: I1205 11:08:13.883455 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tqtr9"] Dec 05 11:08:13.894355 master-0 kubenswrapper[24928]: I1205 11:08:13.894280 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-8tsgs" podStartSLOduration=3.89425898 podStartE2EDuration="3.89425898s" podCreationTimestamp="2025-12-05 11:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:13.892373944 +0000 UTC m=+1253.895567795" watchObservedRunningTime="2025-12-05 11:08:13.89425898 +0000 UTC m=+1253.897452841" Dec 05 11:08:13.907224 master-0 kubenswrapper[24928]: W1205 11:08:13.907153 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdfae6e2_22a3_4218_b6e9_2558a61bcead.slice/crio-234f4bd7084a59e3bc7c456b448955c17bdb96d309d8f8b3d624df73cb6026bf WatchSource:0}: Error finding container 234f4bd7084a59e3bc7c456b448955c17bdb96d309d8f8b3d624df73cb6026bf: Status 404 returned error can't find the container with id 234f4bd7084a59e3bc7c456b448955c17bdb96d309d8f8b3d624df73cb6026bf Dec 05 11:08:13.912821 master-0 kubenswrapper[24928]: I1205 11:08:13.912762 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"923a2187-d634-44ba-bb89-bd632b6e2908","Type":"ContainerStarted","Data":"1883ca908e07a5ad552278d8a5ef796feb8a5eac5cbe069929edbd04b5a61254"} Dec 05 11:08:13.925275 master-0 kubenswrapper[24928]: I1205 11:08:13.924862 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62fbcd8d-b895-4135-bc9a-339f4369e359","Type":"ContainerStarted","Data":"0bd75a6069efdd0216a63272838195c490886efb465319fb9739be80cbcd30d9"} Dec 05 11:08:13.926783 master-0 kubenswrapper[24928]: I1205 11:08:13.926732 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3391ea32-ded0-4fc0-92b6-823023c6b9ca","Type":"ContainerStarted","Data":"2d3da2ece291125b050462d37c95c14783df177c1f5a7cbae9fbd3427ffbdd48"} Dec 05 11:08:14.610998 master-0 kubenswrapper[24928]: I1205 11:08:14.610932 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76dc4d7685-49kgr"] Dec 05 11:08:14.667707 master-0 kubenswrapper[24928]: W1205 11:08:14.667624 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e744111_e3f8_4e33_8fc5_eb86882de858.slice/crio-e5f82097a45f0ed340361d0c4d462d2694a3662c765b49ba8333e54f96d927d5 WatchSource:0}: Error finding container e5f82097a45f0ed340361d0c4d462d2694a3662c765b49ba8333e54f96d927d5: Status 404 returned error can't find the container with id e5f82097a45f0ed340361d0c4d462d2694a3662c765b49ba8333e54f96d927d5 Dec 05 11:08:14.955448 master-0 kubenswrapper[24928]: I1205 11:08:14.952870 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af","Type":"ContainerStarted","Data":"e1a684b73aae7d3f7233bd2f67ab4e51a71fdc242552c5e224aaded0e9aa2b29"} Dec 05 11:08:14.985131 master-0 kubenswrapper[24928]: I1205 11:08:14.985041 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tqtr9" event={"ID":"cdfae6e2-22a3-4218-b6e9-2558a61bcead","Type":"ContainerStarted","Data":"00fff1acca38683d164302d064fbbf231b93dff900bc630d9a907d5c7c4b4e4d"} Dec 05 11:08:14.985131 master-0 kubenswrapper[24928]: I1205 11:08:14.985121 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tqtr9" event={"ID":"cdfae6e2-22a3-4218-b6e9-2558a61bcead","Type":"ContainerStarted","Data":"234f4bd7084a59e3bc7c456b448955c17bdb96d309d8f8b3d624df73cb6026bf"} Dec 05 11:08:14.996446 master-0 kubenswrapper[24928]: I1205 11:08:14.993215 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" event={"ID":"9e744111-e3f8-4e33-8fc5-eb86882de858","Type":"ContainerStarted","Data":"e5f82097a45f0ed340361d0c4d462d2694a3662c765b49ba8333e54f96d927d5"} Dec 05 11:08:15.048054 master-0 kubenswrapper[24928]: I1205 11:08:15.039682 24928 generic.go:334] "Generic (PLEG): container finished" podID="2d8079b6-0de1-4ffd-b008-9d74bf9673ae" containerID="92d2c945339184594773d810897a8f3f1eab7fe95efacf43537287c8391cd24a" exitCode=0 Dec 05 11:08:15.048054 master-0 kubenswrapper[24928]: I1205 11:08:15.039924 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"2d8079b6-0de1-4ffd-b008-9d74bf9673ae","Type":"ContainerDied","Data":"92d2c945339184594773d810897a8f3f1eab7fe95efacf43537287c8391cd24a"} Dec 05 11:08:15.048054 master-0 kubenswrapper[24928]: I1205 11:08:15.044503 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-tqtr9" podStartSLOduration=3.044474993 podStartE2EDuration="3.044474993s" podCreationTimestamp="2025-12-05 11:08:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:15.0304018 +0000 UTC m=+1255.033595661" watchObservedRunningTime="2025-12-05 11:08:15.044474993 +0000 UTC m=+1255.047668844" Dec 05 11:08:16.057001 master-0 kubenswrapper[24928]: I1205 11:08:16.056941 24928 generic.go:334] "Generic (PLEG): container finished" podID="9e744111-e3f8-4e33-8fc5-eb86882de858" containerID="d20dcecce2abb084fb2107251949a872ca38702bf882d2086a76918214f0de4a" exitCode=0 Dec 05 11:08:16.057703 master-0 kubenswrapper[24928]: I1205 11:08:16.057513 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" event={"ID":"9e744111-e3f8-4e33-8fc5-eb86882de858","Type":"ContainerDied","Data":"d20dcecce2abb084fb2107251949a872ca38702bf882d2086a76918214f0de4a"} Dec 05 11:08:16.064245 master-0 kubenswrapper[24928]: I1205 11:08:16.064174 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"2d8079b6-0de1-4ffd-b008-9d74bf9673ae","Type":"ContainerStarted","Data":"bc52717b4a3073e0edb14a1fa4a2724cf45832114b4dad3b49a9392b1b176d56"} Dec 05 11:08:18.859651 master-0 kubenswrapper[24928]: I1205 11:08:18.859584 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:08:18.898439 master-0 kubenswrapper[24928]: I1205 11:08:18.898354 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:19.112013 master-0 kubenswrapper[24928]: I1205 11:08:19.111902 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3391ea32-ded0-4fc0-92b6-823023c6b9ca","Type":"ContainerStarted","Data":"c1b0c05ec78af49de9920697e85ca8bfd7102622bc81b3844ee301f719d705cf"} Dec 05 11:08:19.121843 master-0 kubenswrapper[24928]: I1205 11:08:19.121779 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"2d8079b6-0de1-4ffd-b008-9d74bf9673ae","Type":"ContainerStarted","Data":"6f8e34e8bbcaf6ce9f6ab3fe0684d096df46650986eb560a7ff6adfe6fe0ea75"} Dec 05 11:08:19.121843 master-0 kubenswrapper[24928]: I1205 11:08:19.121846 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"2d8079b6-0de1-4ffd-b008-9d74bf9673ae","Type":"ContainerStarted","Data":"c0981b2098b40476beaf6a4c107c876209dd87cef60125d2b34385c8b6411ebe"} Dec 05 11:08:19.122136 master-0 kubenswrapper[24928]: I1205 11:08:19.122029 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-conductor-0" Dec 05 11:08:19.125511 master-0 kubenswrapper[24928]: I1205 11:08:19.125452 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af","Type":"ContainerStarted","Data":"c1f3ca1dcbb3c749d9e7f2d1d68f311e74b529e2ae8c0597171ed15f354ab507"} Dec 05 11:08:19.125511 master-0 kubenswrapper[24928]: I1205 11:08:19.125490 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af","Type":"ContainerStarted","Data":"0c6a767bf50ca93bc1edb4105e71f9fbf120850d4c709b8e8cf3474d09c72d99"} Dec 05 11:08:19.129655 master-0 kubenswrapper[24928]: I1205 11:08:19.128107 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"923a2187-d634-44ba-bb89-bd632b6e2908","Type":"ContainerStarted","Data":"5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab"} Dec 05 11:08:19.132645 master-0 kubenswrapper[24928]: I1205 11:08:19.132574 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62fbcd8d-b895-4135-bc9a-339f4369e359","Type":"ContainerStarted","Data":"bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247"} Dec 05 11:08:19.132645 master-0 kubenswrapper[24928]: I1205 11:08:19.132639 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62fbcd8d-b895-4135-bc9a-339f4369e359","Type":"ContainerStarted","Data":"5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d"} Dec 05 11:08:19.135063 master-0 kubenswrapper[24928]: I1205 11:08:19.135013 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" event={"ID":"9e744111-e3f8-4e33-8fc5-eb86882de858","Type":"ContainerStarted","Data":"257c7d15b34cdf394b87a4c3e5275102b5afe40f2ad9d00c6788a5b1197ec38f"} Dec 05 11:08:19.135333 master-0 kubenswrapper[24928]: I1205 11:08:19.135294 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:19.209680 master-0 kubenswrapper[24928]: I1205 11:08:19.208357 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=4.05295367 podStartE2EDuration="8.208334819s" podCreationTimestamp="2025-12-05 11:08:11 +0000 UTC" firstStartedPulling="2025-12-05 11:08:13.459366243 +0000 UTC m=+1253.462560094" lastFinishedPulling="2025-12-05 11:08:17.614747392 +0000 UTC m=+1257.617941243" observedRunningTime="2025-12-05 11:08:19.170575398 +0000 UTC m=+1259.173769259" watchObservedRunningTime="2025-12-05 11:08:19.208334819 +0000 UTC m=+1259.211528680" Dec 05 11:08:19.215459 master-0 kubenswrapper[24928]: I1205 11:08:19.215348 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.208822371 podStartE2EDuration="8.215320389s" podCreationTimestamp="2025-12-05 11:08:11 +0000 UTC" firstStartedPulling="2025-12-05 11:08:13.606937502 +0000 UTC m=+1253.610131353" lastFinishedPulling="2025-12-05 11:08:17.61343552 +0000 UTC m=+1257.616629371" observedRunningTime="2025-12-05 11:08:19.189990501 +0000 UTC m=+1259.193184372" watchObservedRunningTime="2025-12-05 11:08:19.215320389 +0000 UTC m=+1259.218514240" Dec 05 11:08:19.236495 master-0 kubenswrapper[24928]: I1205 11:08:19.236328 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.462982821 podStartE2EDuration="8.236301611s" podCreationTimestamp="2025-12-05 11:08:11 +0000 UTC" firstStartedPulling="2025-12-05 11:08:13.842026536 +0000 UTC m=+1253.845220387" lastFinishedPulling="2025-12-05 11:08:17.615345326 +0000 UTC m=+1257.618539177" observedRunningTime="2025-12-05 11:08:19.227786823 +0000 UTC m=+1259.230980764" watchObservedRunningTime="2025-12-05 11:08:19.236301611 +0000 UTC m=+1259.239495462" Dec 05 11:08:19.278445 master-0 kubenswrapper[24928]: I1205 11:08:19.275624 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" podStartSLOduration=7.275598369 podStartE2EDuration="7.275598369s" podCreationTimestamp="2025-12-05 11:08:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:19.260926851 +0000 UTC m=+1259.264120722" watchObservedRunningTime="2025-12-05 11:08:19.275598369 +0000 UTC m=+1259.278792220" Dec 05 11:08:19.329200 master-0 kubenswrapper[24928]: I1205 11:08:19.329091 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-conductor-0" podStartSLOduration=62.002599218 podStartE2EDuration="1m43.329067513s" podCreationTimestamp="2025-12-05 11:06:36 +0000 UTC" firstStartedPulling="2025-12-05 11:06:45.651350703 +0000 UTC m=+1165.654544554" lastFinishedPulling="2025-12-05 11:07:26.977818998 +0000 UTC m=+1206.981012849" observedRunningTime="2025-12-05 11:08:19.305189571 +0000 UTC m=+1259.308383432" watchObservedRunningTime="2025-12-05 11:08:19.329067513 +0000 UTC m=+1259.332261364" Dec 05 11:08:19.345241 master-0 kubenswrapper[24928]: I1205 11:08:19.345107 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.389596221 podStartE2EDuration="8.345088344s" podCreationTimestamp="2025-12-05 11:08:11 +0000 UTC" firstStartedPulling="2025-12-05 11:08:13.659655238 +0000 UTC m=+1253.662849089" lastFinishedPulling="2025-12-05 11:08:17.615147361 +0000 UTC m=+1257.618341212" observedRunningTime="2025-12-05 11:08:19.327612827 +0000 UTC m=+1259.330806688" watchObservedRunningTime="2025-12-05 11:08:19.345088344 +0000 UTC m=+1259.348282195" Dec 05 11:08:20.148260 master-0 kubenswrapper[24928]: I1205 11:08:20.148136 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3391ea32-ded0-4fc0-92b6-823023c6b9ca" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://c1b0c05ec78af49de9920697e85ca8bfd7102622bc81b3844ee301f719d705cf" gracePeriod=30 Dec 05 11:08:20.149529 master-0 kubenswrapper[24928]: I1205 11:08:20.149339 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" containerName="nova-metadata-log" containerID="cri-o://0c6a767bf50ca93bc1edb4105e71f9fbf120850d4c709b8e8cf3474d09c72d99" gracePeriod=30 Dec 05 11:08:20.149529 master-0 kubenswrapper[24928]: I1205 11:08:20.149463 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" containerName="nova-metadata-metadata" containerID="cri-o://c1f3ca1dcbb3c749d9e7f2d1d68f311e74b529e2ae8c0597171ed15f354ab507" gracePeriod=30 Dec 05 11:08:20.149990 master-0 kubenswrapper[24928]: I1205 11:08:20.149935 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-conductor-0" Dec 05 11:08:20.149990 master-0 kubenswrapper[24928]: I1205 11:08:20.149989 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-conductor-0" Dec 05 11:08:20.186604 master-0 kubenswrapper[24928]: I1205 11:08:20.186152 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-conductor-0" Dec 05 11:08:21.184567 master-0 kubenswrapper[24928]: I1205 11:08:21.184500 24928 generic.go:334] "Generic (PLEG): container finished" podID="3391ea32-ded0-4fc0-92b6-823023c6b9ca" containerID="c1b0c05ec78af49de9920697e85ca8bfd7102622bc81b3844ee301f719d705cf" exitCode=0 Dec 05 11:08:21.185187 master-0 kubenswrapper[24928]: I1205 11:08:21.184593 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3391ea32-ded0-4fc0-92b6-823023c6b9ca","Type":"ContainerDied","Data":"c1b0c05ec78af49de9920697e85ca8bfd7102622bc81b3844ee301f719d705cf"} Dec 05 11:08:21.187535 master-0 kubenswrapper[24928]: I1205 11:08:21.187467 24928 generic.go:334] "Generic (PLEG): container finished" podID="cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" containerID="c1f3ca1dcbb3c749d9e7f2d1d68f311e74b529e2ae8c0597171ed15f354ab507" exitCode=0 Dec 05 11:08:21.187627 master-0 kubenswrapper[24928]: I1205 11:08:21.187537 24928 generic.go:334] "Generic (PLEG): container finished" podID="cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" containerID="0c6a767bf50ca93bc1edb4105e71f9fbf120850d4c709b8e8cf3474d09c72d99" exitCode=143 Dec 05 11:08:21.187627 master-0 kubenswrapper[24928]: I1205 11:08:21.187570 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af","Type":"ContainerDied","Data":"c1f3ca1dcbb3c749d9e7f2d1d68f311e74b529e2ae8c0597171ed15f354ab507"} Dec 05 11:08:21.187712 master-0 kubenswrapper[24928]: I1205 11:08:21.187635 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af","Type":"ContainerDied","Data":"0c6a767bf50ca93bc1edb4105e71f9fbf120850d4c709b8e8cf3474d09c72d99"} Dec 05 11:08:21.230089 master-0 kubenswrapper[24928]: I1205 11:08:21.230033 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-conductor-0" Dec 05 11:08:22.015596 master-0 kubenswrapper[24928]: I1205 11:08:22.015531 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:08:22.015596 master-0 kubenswrapper[24928]: I1205 11:08:22.015608 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:08:22.018878 master-0 kubenswrapper[24928]: I1205 11:08:22.018827 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:22.028838 master-0 kubenswrapper[24928]: I1205 11:08:22.028633 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 11:08:22.028838 master-0 kubenswrapper[24928]: I1205 11:08:22.028699 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 11:08:22.073998 master-0 kubenswrapper[24928]: I1205 11:08:22.073921 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 11:08:22.201604 master-0 kubenswrapper[24928]: I1205 11:08:22.201386 24928 generic.go:334] "Generic (PLEG): container finished" podID="f6dde2d1-2c98-48c1-ac60-02129e7543b3" containerID="fa1dbb07b11ef157ddc386720c998f0bd948fbc44b183a8e3569217c865c86c9" exitCode=0 Dec 05 11:08:22.201604 master-0 kubenswrapper[24928]: I1205 11:08:22.201458 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8tsgs" event={"ID":"f6dde2d1-2c98-48c1-ac60-02129e7543b3","Type":"ContainerDied","Data":"fa1dbb07b11ef157ddc386720c998f0bd948fbc44b183a8e3569217c865c86c9"} Dec 05 11:08:22.239152 master-0 kubenswrapper[24928]: I1205 11:08:22.239094 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 11:08:22.824229 master-0 kubenswrapper[24928]: I1205 11:08:22.824168 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 11:08:22.824229 master-0 kubenswrapper[24928]: I1205 11:08:22.824233 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 11:08:23.099858 master-0 kubenswrapper[24928]: I1205 11:08:23.099714 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.128.0.254:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 11:08:23.100150 master-0 kubenswrapper[24928]: I1205 11:08:23.099722 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.128.0.254:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 11:08:23.216335 master-0 kubenswrapper[24928]: I1205 11:08:23.216240 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-conductor-0" Dec 05 11:08:23.420581 master-0 kubenswrapper[24928]: I1205 11:08:23.419628 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:29.801781 master-0 kubenswrapper[24928]: I1205 11:08:29.801718 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:30.312566 master-0 kubenswrapper[24928]: I1205 11:08:30.312496 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:30.331241 master-0 kubenswrapper[24928]: I1205 11:08:30.331167 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:30.331530 master-0 kubenswrapper[24928]: I1205 11:08:30.331263 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af","Type":"ContainerDied","Data":"e1a684b73aae7d3f7233bd2f67ab4e51a71fdc242552c5e224aaded0e9aa2b29"} Dec 05 11:08:30.331530 master-0 kubenswrapper[24928]: I1205 11:08:30.331310 24928 scope.go:117] "RemoveContainer" containerID="c1f3ca1dcbb3c749d9e7f2d1d68f311e74b529e2ae8c0597171ed15f354ab507" Dec 05 11:08:30.332726 master-0 kubenswrapper[24928]: I1205 11:08:30.332691 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-8tsgs" event={"ID":"f6dde2d1-2c98-48c1-ac60-02129e7543b3","Type":"ContainerDied","Data":"483d9bcba42e40d48f087889b4fb4d5b0dafc97bd1a17b36ac01345aec501910"} Dec 05 11:08:30.332726 master-0 kubenswrapper[24928]: I1205 11:08:30.332720 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="483d9bcba42e40d48f087889b4fb4d5b0dafc97bd1a17b36ac01345aec501910" Dec 05 11:08:30.332907 master-0 kubenswrapper[24928]: I1205 11:08:30.332771 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-8tsgs" Dec 05 11:08:30.335881 master-0 kubenswrapper[24928]: I1205 11:08:30.335840 24928 generic.go:334] "Generic (PLEG): container finished" podID="cdfae6e2-22a3-4218-b6e9-2558a61bcead" containerID="00fff1acca38683d164302d064fbbf231b93dff900bc630d9a907d5c7c4b4e4d" exitCode=0 Dec 05 11:08:30.336029 master-0 kubenswrapper[24928]: I1205 11:08:30.335912 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tqtr9" event={"ID":"cdfae6e2-22a3-4218-b6e9-2558a61bcead","Type":"ContainerDied","Data":"00fff1acca38683d164302d064fbbf231b93dff900bc630d9a907d5c7c4b4e4d"} Dec 05 11:08:30.338444 master-0 kubenswrapper[24928]: I1205 11:08:30.338378 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3391ea32-ded0-4fc0-92b6-823023c6b9ca","Type":"ContainerDied","Data":"2d3da2ece291125b050462d37c95c14783df177c1f5a7cbae9fbd3427ffbdd48"} Dec 05 11:08:30.338444 master-0 kubenswrapper[24928]: I1205 11:08:30.338436 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:30.361300 master-0 kubenswrapper[24928]: I1205 11:08:30.361209 24928 scope.go:117] "RemoveContainer" containerID="0c6a767bf50ca93bc1edb4105e71f9fbf120850d4c709b8e8cf3474d09c72d99" Dec 05 11:08:30.405636 master-0 kubenswrapper[24928]: I1205 11:08:30.405564 24928 scope.go:117] "RemoveContainer" containerID="c1b0c05ec78af49de9920697e85ca8bfd7102622bc81b3844ee301f719d705cf" Dec 05 11:08:31.094784 master-0 kubenswrapper[24928]: I1205 11:08:31.094706 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-combined-ca-bundle\") pod \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " Dec 05 11:08:31.094784 master-0 kubenswrapper[24928]: I1205 11:08:31.094786 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-config-data\") pod \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " Dec 05 11:08:31.095337 master-0 kubenswrapper[24928]: I1205 11:08:31.094815 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-logs\") pod \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " Dec 05 11:08:31.095337 master-0 kubenswrapper[24928]: I1205 11:08:31.094849 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-config-data\") pod \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " Dec 05 11:08:31.095337 master-0 kubenswrapper[24928]: I1205 11:08:31.094879 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj8kf\" (UniqueName: \"kubernetes.io/projected/f6dde2d1-2c98-48c1-ac60-02129e7543b3-kube-api-access-hj8kf\") pod \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " Dec 05 11:08:31.095337 master-0 kubenswrapper[24928]: I1205 11:08:31.094958 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-combined-ca-bundle\") pod \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " Dec 05 11:08:31.095337 master-0 kubenswrapper[24928]: I1205 11:08:31.095048 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-combined-ca-bundle\") pod \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " Dec 05 11:08:31.095337 master-0 kubenswrapper[24928]: I1205 11:08:31.095107 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-scripts\") pod \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\" (UID: \"f6dde2d1-2c98-48c1-ac60-02129e7543b3\") " Dec 05 11:08:31.095337 master-0 kubenswrapper[24928]: I1205 11:08:31.095161 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-logs" (OuterVolumeSpecName: "logs") pod "cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" (UID: "cc3ac2a8-410b-40b6-870f-4d9b94c2e1af"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:31.095652 master-0 kubenswrapper[24928]: I1205 11:08:31.095606 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vf7t\" (UniqueName: \"kubernetes.io/projected/3391ea32-ded0-4fc0-92b6-823023c6b9ca-kube-api-access-4vf7t\") pod \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\" (UID: \"3391ea32-ded0-4fc0-92b6-823023c6b9ca\") " Dec 05 11:08:31.095701 master-0 kubenswrapper[24928]: I1205 11:08:31.095686 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-config-data\") pod \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " Dec 05 11:08:31.095736 master-0 kubenswrapper[24928]: I1205 11:08:31.095715 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rrmg\" (UniqueName: \"kubernetes.io/projected/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-kube-api-access-4rrmg\") pod \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\" (UID: \"cc3ac2a8-410b-40b6-870f-4d9b94c2e1af\") " Dec 05 11:08:31.100067 master-0 kubenswrapper[24928]: I1205 11:08:31.097873 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-scripts" (OuterVolumeSpecName: "scripts") pod "f6dde2d1-2c98-48c1-ac60-02129e7543b3" (UID: "f6dde2d1-2c98-48c1-ac60-02129e7543b3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:31.100067 master-0 kubenswrapper[24928]: I1205 11:08:31.099115 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:31.100067 master-0 kubenswrapper[24928]: I1205 11:08:31.099577 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:31.101994 master-0 kubenswrapper[24928]: I1205 11:08:31.101278 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6dde2d1-2c98-48c1-ac60-02129e7543b3-kube-api-access-hj8kf" (OuterVolumeSpecName: "kube-api-access-hj8kf") pod "f6dde2d1-2c98-48c1-ac60-02129e7543b3" (UID: "f6dde2d1-2c98-48c1-ac60-02129e7543b3"). InnerVolumeSpecName "kube-api-access-hj8kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:31.101994 master-0 kubenswrapper[24928]: I1205 11:08:31.101814 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-kube-api-access-4rrmg" (OuterVolumeSpecName: "kube-api-access-4rrmg") pod "cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" (UID: "cc3ac2a8-410b-40b6-870f-4d9b94c2e1af"). InnerVolumeSpecName "kube-api-access-4rrmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:31.111019 master-0 kubenswrapper[24928]: I1205 11:08:31.105562 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3391ea32-ded0-4fc0-92b6-823023c6b9ca-kube-api-access-4vf7t" (OuterVolumeSpecName: "kube-api-access-4vf7t") pod "3391ea32-ded0-4fc0-92b6-823023c6b9ca" (UID: "3391ea32-ded0-4fc0-92b6-823023c6b9ca"). InnerVolumeSpecName "kube-api-access-4vf7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:31.128621 master-0 kubenswrapper[24928]: I1205 11:08:31.128536 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-config-data" (OuterVolumeSpecName: "config-data") pod "3391ea32-ded0-4fc0-92b6-823023c6b9ca" (UID: "3391ea32-ded0-4fc0-92b6-823023c6b9ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:31.135033 master-0 kubenswrapper[24928]: I1205 11:08:31.134978 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" (UID: "cc3ac2a8-410b-40b6-870f-4d9b94c2e1af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:31.135610 master-0 kubenswrapper[24928]: I1205 11:08:31.135553 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-config-data" (OuterVolumeSpecName: "config-data") pod "f6dde2d1-2c98-48c1-ac60-02129e7543b3" (UID: "f6dde2d1-2c98-48c1-ac60-02129e7543b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:31.142339 master-0 kubenswrapper[24928]: I1205 11:08:31.142254 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3391ea32-ded0-4fc0-92b6-823023c6b9ca" (UID: "3391ea32-ded0-4fc0-92b6-823023c6b9ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:31.149197 master-0 kubenswrapper[24928]: I1205 11:08:31.149088 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-config-data" (OuterVolumeSpecName: "config-data") pod "cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" (UID: "cc3ac2a8-410b-40b6-870f-4d9b94c2e1af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:31.150812 master-0 kubenswrapper[24928]: I1205 11:08:31.150751 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6dde2d1-2c98-48c1-ac60-02129e7543b3" (UID: "f6dde2d1-2c98-48c1-ac60-02129e7543b3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:31.208458 master-0 kubenswrapper[24928]: I1205 11:08:31.204595 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:31.208458 master-0 kubenswrapper[24928]: I1205 11:08:31.204638 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rrmg\" (UniqueName: \"kubernetes.io/projected/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-kube-api-access-4rrmg\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:31.208458 master-0 kubenswrapper[24928]: I1205 11:08:31.204655 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:31.208458 master-0 kubenswrapper[24928]: I1205 11:08:31.206616 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:31.208458 master-0 kubenswrapper[24928]: I1205 11:08:31.206665 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:31.208458 master-0 kubenswrapper[24928]: I1205 11:08:31.206683 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj8kf\" (UniqueName: \"kubernetes.io/projected/f6dde2d1-2c98-48c1-ac60-02129e7543b3-kube-api-access-hj8kf\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:31.208458 master-0 kubenswrapper[24928]: I1205 11:08:31.206696 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3391ea32-ded0-4fc0-92b6-823023c6b9ca-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:31.208458 master-0 kubenswrapper[24928]: I1205 11:08:31.206707 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6dde2d1-2c98-48c1-ac60-02129e7543b3-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:31.208458 master-0 kubenswrapper[24928]: I1205 11:08:31.206718 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vf7t\" (UniqueName: \"kubernetes.io/projected/3391ea32-ded0-4fc0-92b6-823023c6b9ca-kube-api-access-4vf7t\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:31.277832 master-0 kubenswrapper[24928]: I1205 11:08:31.275618 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68cb7b79bf-rs4hj"] Dec 05 11:08:31.277832 master-0 kubenswrapper[24928]: I1205 11:08:31.275918 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" podUID="b7a38786-cb62-46b5-8a00-54f6d01545b8" containerName="dnsmasq-dns" containerID="cri-o://6d1b122f166fe65e8e7588f3cd14d0233e19f369beb8056737fd38fc24cd8f3a" gracePeriod=10 Dec 05 11:08:31.410824 master-0 kubenswrapper[24928]: I1205 11:08:31.402645 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:31.486287 master-0 kubenswrapper[24928]: I1205 11:08:31.486201 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:08:31.502529 master-0 kubenswrapper[24928]: I1205 11:08:31.501015 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:08:31.560654 master-0 kubenswrapper[24928]: I1205 11:08:31.560601 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: E1205 11:08:31.561437 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" containerName="nova-metadata-log" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: I1205 11:08:31.561466 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" containerName="nova-metadata-log" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: E1205 11:08:31.561499 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6dde2d1-2c98-48c1-ac60-02129e7543b3" containerName="nova-manage" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: I1205 11:08:31.561509 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6dde2d1-2c98-48c1-ac60-02129e7543b3" containerName="nova-manage" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: E1205 11:08:31.561539 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3391ea32-ded0-4fc0-92b6-823023c6b9ca" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: I1205 11:08:31.561549 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="3391ea32-ded0-4fc0-92b6-823023c6b9ca" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: E1205 11:08:31.561579 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" containerName="nova-metadata-metadata" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: I1205 11:08:31.561588 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" containerName="nova-metadata-metadata" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: I1205 11:08:31.561874 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="3391ea32-ded0-4fc0-92b6-823023c6b9ca" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: I1205 11:08:31.561919 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" containerName="nova-metadata-log" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: I1205 11:08:31.561968 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6dde2d1-2c98-48c1-ac60-02129e7543b3" containerName="nova-manage" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: I1205 11:08:31.561984 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" containerName="nova-metadata-metadata" Dec 05 11:08:31.567507 master-0 kubenswrapper[24928]: I1205 11:08:31.563000 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.588261 master-0 kubenswrapper[24928]: I1205 11:08:31.588213 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 11:08:31.588474 master-0 kubenswrapper[24928]: I1205 11:08:31.588453 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 05 11:08:31.588656 master-0 kubenswrapper[24928]: I1205 11:08:31.588619 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:08:31.588842 master-0 kubenswrapper[24928]: I1205 11:08:31.588816 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 05 11:08:31.602903 master-0 kubenswrapper[24928]: I1205 11:08:31.602855 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:31.623573 master-0 kubenswrapper[24928]: I1205 11:08:31.623268 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:31.639782 master-0 kubenswrapper[24928]: I1205 11:08:31.639729 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:31.671267 master-0 kubenswrapper[24928]: I1205 11:08:31.671135 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:31.671500 master-0 kubenswrapper[24928]: I1205 11:08:31.671266 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:31.674029 master-0 kubenswrapper[24928]: I1205 11:08:31.673999 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 11:08:31.675612 master-0 kubenswrapper[24928]: I1205 11:08:31.675569 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 11:08:31.741258 master-0 kubenswrapper[24928]: I1205 11:08:31.741184 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.741258 master-0 kubenswrapper[24928]: I1205 11:08:31.741281 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.741855 master-0 kubenswrapper[24928]: I1205 11:08:31.741671 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.741931 master-0 kubenswrapper[24928]: I1205 11:08:31.741908 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.742470 master-0 kubenswrapper[24928]: I1205 11:08:31.742287 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhnp2\" (UniqueName: \"kubernetes.io/projected/20748170-9dfa-4816-97cd-f6a12aaf4531-kube-api-access-fhnp2\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.856798 master-0 kubenswrapper[24928]: I1205 11:08:31.855129 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.856798 master-0 kubenswrapper[24928]: I1205 11:08:31.855253 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.856798 master-0 kubenswrapper[24928]: I1205 11:08:31.855387 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-config-data\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.856798 master-0 kubenswrapper[24928]: I1205 11:08:31.855510 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhnp2\" (UniqueName: \"kubernetes.io/projected/20748170-9dfa-4816-97cd-f6a12aaf4531-kube-api-access-fhnp2\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.856798 master-0 kubenswrapper[24928]: I1205 11:08:31.855550 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.856798 master-0 kubenswrapper[24928]: I1205 11:08:31.855582 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.856798 master-0 kubenswrapper[24928]: I1205 11:08:31.855624 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxb95\" (UniqueName: \"kubernetes.io/projected/83110f63-0a22-47c7-a52c-0bab96c66f54-kube-api-access-zxb95\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.856798 master-0 kubenswrapper[24928]: I1205 11:08:31.855702 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.856798 master-0 kubenswrapper[24928]: I1205 11:08:31.855762 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.856798 master-0 kubenswrapper[24928]: I1205 11:08:31.855786 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83110f63-0a22-47c7-a52c-0bab96c66f54-logs\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.861433 master-0 kubenswrapper[24928]: I1205 11:08:31.861353 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.892252 master-0 kubenswrapper[24928]: I1205 11:08:31.892191 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.892492 master-0 kubenswrapper[24928]: I1205 11:08:31.892317 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.893836 master-0 kubenswrapper[24928]: I1205 11:08:31.893642 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20748170-9dfa-4816-97cd-f6a12aaf4531-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.908110 master-0 kubenswrapper[24928]: I1205 11:08:31.908047 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhnp2\" (UniqueName: \"kubernetes.io/projected/20748170-9dfa-4816-97cd-f6a12aaf4531-kube-api-access-fhnp2\") pod \"nova-cell1-novncproxy-0\" (UID: \"20748170-9dfa-4816-97cd-f6a12aaf4531\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:31.958444 master-0 kubenswrapper[24928]: I1205 11:08:31.958280 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.958444 master-0 kubenswrapper[24928]: I1205 11:08:31.958334 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.958444 master-0 kubenswrapper[24928]: I1205 11:08:31.958359 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxb95\" (UniqueName: \"kubernetes.io/projected/83110f63-0a22-47c7-a52c-0bab96c66f54-kube-api-access-zxb95\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.958735 master-0 kubenswrapper[24928]: I1205 11:08:31.958570 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83110f63-0a22-47c7-a52c-0bab96c66f54-logs\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.960761 master-0 kubenswrapper[24928]: I1205 11:08:31.959213 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83110f63-0a22-47c7-a52c-0bab96c66f54-logs\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.960761 master-0 kubenswrapper[24928]: I1205 11:08:31.959355 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-config-data\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.964004 master-0 kubenswrapper[24928]: I1205 11:08:31.962614 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.964004 master-0 kubenswrapper[24928]: I1205 11:08:31.963509 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-config-data\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.964004 master-0 kubenswrapper[24928]: I1205 11:08:31.964001 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:31.985541 master-0 kubenswrapper[24928]: I1205 11:08:31.984057 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxb95\" (UniqueName: \"kubernetes.io/projected/83110f63-0a22-47c7-a52c-0bab96c66f54-kube-api-access-zxb95\") pod \"nova-metadata-0\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " pod="openstack/nova-metadata-0" Dec 05 11:08:32.005274 master-0 kubenswrapper[24928]: I1205 11:08:32.005051 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:32.014672 master-0 kubenswrapper[24928]: I1205 11:08:32.013334 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:32.014672 master-0 kubenswrapper[24928]: I1205 11:08:32.013902 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:32.020543 master-0 kubenswrapper[24928]: I1205 11:08:32.020402 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 11:08:32.020962 master-0 kubenswrapper[24928]: I1205 11:08:32.020870 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 11:08:32.024666 master-0 kubenswrapper[24928]: I1205 11:08:32.023608 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 11:08:32.028456 master-0 kubenswrapper[24928]: I1205 11:08:32.028063 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 11:08:32.202463 master-0 kubenswrapper[24928]: I1205 11:08:32.196893 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kd7f\" (UniqueName: \"kubernetes.io/projected/cdfae6e2-22a3-4218-b6e9-2558a61bcead-kube-api-access-8kd7f\") pod \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " Dec 05 11:08:32.202463 master-0 kubenswrapper[24928]: I1205 11:08:32.197051 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-combined-ca-bundle\") pod \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " Dec 05 11:08:32.202463 master-0 kubenswrapper[24928]: I1205 11:08:32.197369 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-scripts\") pod \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " Dec 05 11:08:32.202463 master-0 kubenswrapper[24928]: I1205 11:08:32.197553 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-config-data\") pod \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\" (UID: \"cdfae6e2-22a3-4218-b6e9-2558a61bcead\") " Dec 05 11:08:32.240785 master-0 kubenswrapper[24928]: I1205 11:08:32.237827 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdfae6e2-22a3-4218-b6e9-2558a61bcead-kube-api-access-8kd7f" (OuterVolumeSpecName: "kube-api-access-8kd7f") pod "cdfae6e2-22a3-4218-b6e9-2558a61bcead" (UID: "cdfae6e2-22a3-4218-b6e9-2558a61bcead"). InnerVolumeSpecName "kube-api-access-8kd7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:32.249449 master-0 kubenswrapper[24928]: I1205 11:08:32.247052 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cdfae6e2-22a3-4218-b6e9-2558a61bcead" (UID: "cdfae6e2-22a3-4218-b6e9-2558a61bcead"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:32.249449 master-0 kubenswrapper[24928]: I1205 11:08:32.247194 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-scripts" (OuterVolumeSpecName: "scripts") pod "cdfae6e2-22a3-4218-b6e9-2558a61bcead" (UID: "cdfae6e2-22a3-4218-b6e9-2558a61bcead"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:32.311454 master-0 kubenswrapper[24928]: I1205 11:08:32.306924 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:32.311454 master-0 kubenswrapper[24928]: I1205 11:08:32.306970 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kd7f\" (UniqueName: \"kubernetes.io/projected/cdfae6e2-22a3-4218-b6e9-2558a61bcead-kube-api-access-8kd7f\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:32.311454 master-0 kubenswrapper[24928]: I1205 11:08:32.306981 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:32.325448 master-0 kubenswrapper[24928]: I1205 11:08:32.323027 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-config-data" (OuterVolumeSpecName: "config-data") pod "cdfae6e2-22a3-4218-b6e9-2558a61bcead" (UID: "cdfae6e2-22a3-4218-b6e9-2558a61bcead"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:32.347157 master-0 kubenswrapper[24928]: I1205 11:08:32.347098 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3391ea32-ded0-4fc0-92b6-823023c6b9ca" path="/var/lib/kubelet/pods/3391ea32-ded0-4fc0-92b6-823023c6b9ca/volumes" Dec 05 11:08:32.368357 master-0 kubenswrapper[24928]: I1205 11:08:32.368292 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc3ac2a8-410b-40b6-870f-4d9b94c2e1af" path="/var/lib/kubelet/pods/cc3ac2a8-410b-40b6-870f-4d9b94c2e1af/volumes" Dec 05 11:08:32.409018 master-0 kubenswrapper[24928]: I1205 11:08:32.408952 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cdfae6e2-22a3-4218-b6e9-2558a61bcead-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:32.469453 master-0 kubenswrapper[24928]: I1205 11:08:32.464860 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-tqtr9" event={"ID":"cdfae6e2-22a3-4218-b6e9-2558a61bcead","Type":"ContainerDied","Data":"234f4bd7084a59e3bc7c456b448955c17bdb96d309d8f8b3d624df73cb6026bf"} Dec 05 11:08:32.469453 master-0 kubenswrapper[24928]: I1205 11:08:32.464910 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="234f4bd7084a59e3bc7c456b448955c17bdb96d309d8f8b3d624df73cb6026bf" Dec 05 11:08:32.469453 master-0 kubenswrapper[24928]: I1205 11:08:32.464982 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-tqtr9" Dec 05 11:08:32.518401 master-0 kubenswrapper[24928]: I1205 11:08:32.510118 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:32.518766 master-0 kubenswrapper[24928]: I1205 11:08:32.518537 24928 generic.go:334] "Generic (PLEG): container finished" podID="b7a38786-cb62-46b5-8a00-54f6d01545b8" containerID="6d1b122f166fe65e8e7588f3cd14d0233e19f369beb8056737fd38fc24cd8f3a" exitCode=0 Dec 05 11:08:32.518766 master-0 kubenswrapper[24928]: I1205 11:08:32.518605 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" event={"ID":"b7a38786-cb62-46b5-8a00-54f6d01545b8","Type":"ContainerDied","Data":"6d1b122f166fe65e8e7588f3cd14d0233e19f369beb8056737fd38fc24cd8f3a"} Dec 05 11:08:32.518766 master-0 kubenswrapper[24928]: I1205 11:08:32.518632 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" event={"ID":"b7a38786-cb62-46b5-8a00-54f6d01545b8","Type":"ContainerDied","Data":"d49450deb614190049128fc15b51a69553abc017a36e43274fc10ae71124834a"} Dec 05 11:08:32.518766 master-0 kubenswrapper[24928]: I1205 11:08:32.518643 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d49450deb614190049128fc15b51a69553abc017a36e43274fc10ae71124834a" Dec 05 11:08:32.518766 master-0 kubenswrapper[24928]: I1205 11:08:32.518763 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: I1205 11:08:32.523765 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: E1205 11:08:32.524550 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7a38786-cb62-46b5-8a00-54f6d01545b8" containerName="dnsmasq-dns" Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: I1205 11:08:32.524572 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7a38786-cb62-46b5-8a00-54f6d01545b8" containerName="dnsmasq-dns" Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: E1205 11:08:32.524625 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdfae6e2-22a3-4218-b6e9-2558a61bcead" containerName="nova-cell1-conductor-db-sync" Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: I1205 11:08:32.524632 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdfae6e2-22a3-4218-b6e9-2558a61bcead" containerName="nova-cell1-conductor-db-sync" Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: E1205 11:08:32.524659 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7a38786-cb62-46b5-8a00-54f6d01545b8" containerName="init" Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: I1205 11:08:32.524666 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7a38786-cb62-46b5-8a00-54f6d01545b8" containerName="init" Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: I1205 11:08:32.524998 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7a38786-cb62-46b5-8a00-54f6d01545b8" containerName="dnsmasq-dns" Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: I1205 11:08:32.525044 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdfae6e2-22a3-4218-b6e9-2558a61bcead" containerName="nova-cell1-conductor-db-sync" Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: I1205 11:08:32.525985 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-compute-ironic-compute-0" event={"ID":"fed98f34-47f9-4239-ac14-637b8ca68b73","Type":"ContainerStarted","Data":"d869df2b4d02d9aa840d494ca3e5e8046a3119ed8abcec70177687d874796ec7"} Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: I1205 11:08:32.526076 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: I1205 11:08:32.526467 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 11:08:32.527414 master-0 kubenswrapper[24928]: I1205 11:08:32.527169 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:32.540462 master-0 kubenswrapper[24928]: I1205 11:08:32.534009 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 11:08:32.560805 master-0 kubenswrapper[24928]: I1205 11:08:32.551016 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 11:08:32.560805 master-0 kubenswrapper[24928]: I1205 11:08:32.555400 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:08:32.560805 master-0 kubenswrapper[24928]: I1205 11:08:32.555745 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="923a2187-d634-44ba-bb89-bd632b6e2908" containerName="nova-scheduler-scheduler" containerID="cri-o://5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab" gracePeriod=30 Dec 05 11:08:32.588729 master-0 kubenswrapper[24928]: I1205 11:08:32.583589 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 11:08:32.609517 master-0 kubenswrapper[24928]: I1205 11:08:32.608482 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:32.633376 master-0 kubenswrapper[24928]: I1205 11:08:32.632412 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-swift-storage-0\") pod \"b7a38786-cb62-46b5-8a00-54f6d01545b8\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " Dec 05 11:08:32.633376 master-0 kubenswrapper[24928]: I1205 11:08:32.632699 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-nb\") pod \"b7a38786-cb62-46b5-8a00-54f6d01545b8\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " Dec 05 11:08:32.633376 master-0 kubenswrapper[24928]: I1205 11:08:32.632783 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmbcg\" (UniqueName: \"kubernetes.io/projected/b7a38786-cb62-46b5-8a00-54f6d01545b8-kube-api-access-hmbcg\") pod \"b7a38786-cb62-46b5-8a00-54f6d01545b8\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " Dec 05 11:08:32.633376 master-0 kubenswrapper[24928]: I1205 11:08:32.632810 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-svc\") pod \"b7a38786-cb62-46b5-8a00-54f6d01545b8\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " Dec 05 11:08:32.633376 master-0 kubenswrapper[24928]: I1205 11:08:32.632882 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-config\") pod \"b7a38786-cb62-46b5-8a00-54f6d01545b8\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " Dec 05 11:08:32.633376 master-0 kubenswrapper[24928]: I1205 11:08:32.632977 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-sb\") pod \"b7a38786-cb62-46b5-8a00-54f6d01545b8\" (UID: \"b7a38786-cb62-46b5-8a00-54f6d01545b8\") " Dec 05 11:08:32.633376 master-0 kubenswrapper[24928]: I1205 11:08:32.633365 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv7hd\" (UniqueName: \"kubernetes.io/projected/1dd6235d-db89-4668-a6ab-510f1aafb764-kube-api-access-sv7hd\") pod \"nova-cell1-conductor-0\" (UID: \"1dd6235d-db89-4668-a6ab-510f1aafb764\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:32.634056 master-0 kubenswrapper[24928]: I1205 11:08:32.633616 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd6235d-db89-4668-a6ab-510f1aafb764-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1dd6235d-db89-4668-a6ab-510f1aafb764\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:32.634056 master-0 kubenswrapper[24928]: I1205 11:08:32.633703 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dd6235d-db89-4668-a6ab-510f1aafb764-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1dd6235d-db89-4668-a6ab-510f1aafb764\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:32.657966 master-0 kubenswrapper[24928]: I1205 11:08:32.654747 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7a38786-cb62-46b5-8a00-54f6d01545b8-kube-api-access-hmbcg" (OuterVolumeSpecName: "kube-api-access-hmbcg") pod "b7a38786-cb62-46b5-8a00-54f6d01545b8" (UID: "b7a38786-cb62-46b5-8a00-54f6d01545b8"). InnerVolumeSpecName "kube-api-access-hmbcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:32.657966 master-0 kubenswrapper[24928]: I1205 11:08:32.657736 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-compute-ironic-compute-0" Dec 05 11:08:32.711465 master-0 kubenswrapper[24928]: I1205 11:08:32.711143 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-compute-ironic-compute-0" podStartSLOduration=2.975242645 podStartE2EDuration="21.711119897s" podCreationTimestamp="2025-12-05 11:08:11 +0000 UTC" firstStartedPulling="2025-12-05 11:08:12.5371324 +0000 UTC m=+1252.540326251" lastFinishedPulling="2025-12-05 11:08:31.273009652 +0000 UTC m=+1271.276203503" observedRunningTime="2025-12-05 11:08:32.615026853 +0000 UTC m=+1272.618220714" watchObservedRunningTime="2025-12-05 11:08:32.711119897 +0000 UTC m=+1272.714313748" Dec 05 11:08:32.726469 master-0 kubenswrapper[24928]: I1205 11:08:32.726180 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b7a38786-cb62-46b5-8a00-54f6d01545b8" (UID: "b7a38786-cb62-46b5-8a00-54f6d01545b8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:32.727648 master-0 kubenswrapper[24928]: I1205 11:08:32.726576 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b7a38786-cb62-46b5-8a00-54f6d01545b8" (UID: "b7a38786-cb62-46b5-8a00-54f6d01545b8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:32.736785 master-0 kubenswrapper[24928]: I1205 11:08:32.734046 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-config" (OuterVolumeSpecName: "config") pod "b7a38786-cb62-46b5-8a00-54f6d01545b8" (UID: "b7a38786-cb62-46b5-8a00-54f6d01545b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:32.747790 master-0 kubenswrapper[24928]: I1205 11:08:32.746955 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd6235d-db89-4668-a6ab-510f1aafb764-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1dd6235d-db89-4668-a6ab-510f1aafb764\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:32.747790 master-0 kubenswrapper[24928]: I1205 11:08:32.747154 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dd6235d-db89-4668-a6ab-510f1aafb764-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1dd6235d-db89-4668-a6ab-510f1aafb764\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:32.747790 master-0 kubenswrapper[24928]: I1205 11:08:32.747390 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv7hd\" (UniqueName: \"kubernetes.io/projected/1dd6235d-db89-4668-a6ab-510f1aafb764-kube-api-access-sv7hd\") pod \"nova-cell1-conductor-0\" (UID: \"1dd6235d-db89-4668-a6ab-510f1aafb764\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:32.748106 master-0 kubenswrapper[24928]: I1205 11:08:32.747914 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmbcg\" (UniqueName: \"kubernetes.io/projected/b7a38786-cb62-46b5-8a00-54f6d01545b8-kube-api-access-hmbcg\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:32.748106 master-0 kubenswrapper[24928]: I1205 11:08:32.747945 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:32.748106 master-0 kubenswrapper[24928]: I1205 11:08:32.747960 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:32.748106 master-0 kubenswrapper[24928]: I1205 11:08:32.747969 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-sb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:32.755619 master-0 kubenswrapper[24928]: I1205 11:08:32.755563 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dd6235d-db89-4668-a6ab-510f1aafb764-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"1dd6235d-db89-4668-a6ab-510f1aafb764\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:32.764114 master-0 kubenswrapper[24928]: I1205 11:08:32.760295 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dd6235d-db89-4668-a6ab-510f1aafb764-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"1dd6235d-db89-4668-a6ab-510f1aafb764\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:32.782139 master-0 kubenswrapper[24928]: I1205 11:08:32.779030 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b7a38786-cb62-46b5-8a00-54f6d01545b8" (UID: "b7a38786-cb62-46b5-8a00-54f6d01545b8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:32.782139 master-0 kubenswrapper[24928]: I1205 11:08:32.780153 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b7a38786-cb62-46b5-8a00-54f6d01545b8" (UID: "b7a38786-cb62-46b5-8a00-54f6d01545b8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:32.783494 master-0 kubenswrapper[24928]: I1205 11:08:32.783456 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv7hd\" (UniqueName: \"kubernetes.io/projected/1dd6235d-db89-4668-a6ab-510f1aafb764-kube-api-access-sv7hd\") pod \"nova-cell1-conductor-0\" (UID: \"1dd6235d-db89-4668-a6ab-510f1aafb764\") " pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:32.862018 master-0 kubenswrapper[24928]: I1205 11:08:32.857771 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:32.862018 master-0 kubenswrapper[24928]: I1205 11:08:32.857856 24928 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b7a38786-cb62-46b5-8a00-54f6d01545b8-dns-swift-storage-0\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:32.878967 master-0 kubenswrapper[24928]: I1205 11:08:32.874742 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:32.902660 master-0 kubenswrapper[24928]: I1205 11:08:32.902113 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c468788f-s9m47"] Dec 05 11:08:32.937533 master-0 kubenswrapper[24928]: I1205 11:08:32.936931 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c468788f-s9m47"] Dec 05 11:08:32.937533 master-0 kubenswrapper[24928]: I1205 11:08:32.937087 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.077048 master-0 kubenswrapper[24928]: I1205 11:08:33.076968 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kndsz\" (UniqueName: \"kubernetes.io/projected/c60d3d04-41e1-4823-b209-18644852fbca-kube-api-access-kndsz\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.077293 master-0 kubenswrapper[24928]: I1205 11:08:33.077081 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-ovsdbserver-sb\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.077293 master-0 kubenswrapper[24928]: I1205 11:08:33.077227 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-dns-svc\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.077404 master-0 kubenswrapper[24928]: I1205 11:08:33.077276 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-dns-swift-storage-0\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.077670 master-0 kubenswrapper[24928]: I1205 11:08:33.077645 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-ovsdbserver-nb\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.077924 master-0 kubenswrapper[24928]: I1205 11:08:33.077757 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-config\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.190692 master-0 kubenswrapper[24928]: I1205 11:08:33.190569 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 11:08:33.194782 master-0 kubenswrapper[24928]: I1205 11:08:33.193909 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-ovsdbserver-nb\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.194782 master-0 kubenswrapper[24928]: I1205 11:08:33.194071 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-config\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.195677 master-0 kubenswrapper[24928]: I1205 11:08:33.194815 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-ovsdbserver-nb\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.195677 master-0 kubenswrapper[24928]: I1205 11:08:33.195392 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-config\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.203383 master-0 kubenswrapper[24928]: I1205 11:08:33.202932 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kndsz\" (UniqueName: \"kubernetes.io/projected/c60d3d04-41e1-4823-b209-18644852fbca-kube-api-access-kndsz\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.203383 master-0 kubenswrapper[24928]: I1205 11:08:33.202993 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-ovsdbserver-sb\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.203383 master-0 kubenswrapper[24928]: I1205 11:08:33.203086 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-dns-svc\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.203383 master-0 kubenswrapper[24928]: I1205 11:08:33.203122 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-dns-swift-storage-0\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.205246 master-0 kubenswrapper[24928]: I1205 11:08:33.204013 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-dns-swift-storage-0\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.205338 master-0 kubenswrapper[24928]: I1205 11:08:33.205246 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-dns-svc\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.235046 master-0 kubenswrapper[24928]: I1205 11:08:33.210917 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c60d3d04-41e1-4823-b209-18644852fbca-ovsdbserver-sb\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.235046 master-0 kubenswrapper[24928]: I1205 11:08:33.233449 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kndsz\" (UniqueName: \"kubernetes.io/projected/c60d3d04-41e1-4823-b209-18644852fbca-kube-api-access-kndsz\") pod \"dnsmasq-dns-57c468788f-s9m47\" (UID: \"c60d3d04-41e1-4823-b209-18644852fbca\") " pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.294286 master-0 kubenswrapper[24928]: I1205 11:08:33.293452 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:33.323462 master-0 kubenswrapper[24928]: I1205 11:08:33.323394 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:33.599090 master-0 kubenswrapper[24928]: W1205 11:08:33.588572 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1dd6235d_db89_4668_a6ab_510f1aafb764.slice/crio-e951f34b434d7776dcdc3eab4b1c4580933003b004d7b38df27e9e1b2256103c WatchSource:0}: Error finding container e951f34b434d7776dcdc3eab4b1c4580933003b004d7b38df27e9e1b2256103c: Status 404 returned error can't find the container with id e951f34b434d7776dcdc3eab4b1c4580933003b004d7b38df27e9e1b2256103c Dec 05 11:08:33.599090 master-0 kubenswrapper[24928]: I1205 11:08:33.597703 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 11:08:33.632492 master-0 kubenswrapper[24928]: I1205 11:08:33.632397 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83110f63-0a22-47c7-a52c-0bab96c66f54","Type":"ContainerStarted","Data":"29ee894fc75f2bab4ac66b5620bccc681d7007519b2c16581cd71a0be3888faa"} Dec 05 11:08:33.653008 master-0 kubenswrapper[24928]: I1205 11:08:33.651568 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"20748170-9dfa-4816-97cd-f6a12aaf4531","Type":"ContainerStarted","Data":"d1626bfa42b2728ce738b427fe0d0617076038faf82fbc6796e407bd12eb7900"} Dec 05 11:08:33.653008 master-0 kubenswrapper[24928]: I1205 11:08:33.651676 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68cb7b79bf-rs4hj" Dec 05 11:08:33.656844 master-0 kubenswrapper[24928]: I1205 11:08:33.656698 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerName="nova-api-log" containerID="cri-o://5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d" gracePeriod=30 Dec 05 11:08:33.657028 master-0 kubenswrapper[24928]: I1205 11:08:33.656996 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerName="nova-api-api" containerID="cri-o://bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247" gracePeriod=30 Dec 05 11:08:33.767899 master-0 kubenswrapper[24928]: I1205 11:08:33.767784 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68cb7b79bf-rs4hj"] Dec 05 11:08:33.784805 master-0 kubenswrapper[24928]: I1205 11:08:33.784736 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68cb7b79bf-rs4hj"] Dec 05 11:08:33.967857 master-0 kubenswrapper[24928]: I1205 11:08:33.967723 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c468788f-s9m47"] Dec 05 11:08:33.979992 master-0 kubenswrapper[24928]: W1205 11:08:33.979940 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc60d3d04_41e1_4823_b209_18644852fbca.slice/crio-3951c9fef19e689b006d0afb1ac5f92b7012b03a375532b3fc2179bbc5ba3245 WatchSource:0}: Error finding container 3951c9fef19e689b006d0afb1ac5f92b7012b03a375532b3fc2179bbc5ba3245: Status 404 returned error can't find the container with id 3951c9fef19e689b006d0afb1ac5f92b7012b03a375532b3fc2179bbc5ba3245 Dec 05 11:08:34.242455 master-0 kubenswrapper[24928]: I1205 11:08:34.240533 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7a38786-cb62-46b5-8a00-54f6d01545b8" path="/var/lib/kubelet/pods/b7a38786-cb62-46b5-8a00-54f6d01545b8/volumes" Dec 05 11:08:34.675505 master-0 kubenswrapper[24928]: I1205 11:08:34.673320 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"20748170-9dfa-4816-97cd-f6a12aaf4531","Type":"ContainerStarted","Data":"c6439fe2e24fb156d6732992da89aebbc8f3c94b0ce5aa85b5edc0abb95ec964"} Dec 05 11:08:34.677406 master-0 kubenswrapper[24928]: I1205 11:08:34.676816 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1dd6235d-db89-4668-a6ab-510f1aafb764","Type":"ContainerStarted","Data":"0352870f1b27c03ffc1ba8b36b19726d8594570c949d3a107a4f688ddcf3daa0"} Dec 05 11:08:34.677406 master-0 kubenswrapper[24928]: I1205 11:08:34.676905 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"1dd6235d-db89-4668-a6ab-510f1aafb764","Type":"ContainerStarted","Data":"e951f34b434d7776dcdc3eab4b1c4580933003b004d7b38df27e9e1b2256103c"} Dec 05 11:08:34.678546 master-0 kubenswrapper[24928]: I1205 11:08:34.678206 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:34.682443 master-0 kubenswrapper[24928]: I1205 11:08:34.681787 24928 generic.go:334] "Generic (PLEG): container finished" podID="c60d3d04-41e1-4823-b209-18644852fbca" containerID="2a3aa65e1ca6f08539e5ed7b278bf25fdf9162670956fe35ca0428223c95cd86" exitCode=0 Dec 05 11:08:34.686437 master-0 kubenswrapper[24928]: I1205 11:08:34.681889 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c468788f-s9m47" event={"ID":"c60d3d04-41e1-4823-b209-18644852fbca","Type":"ContainerDied","Data":"2a3aa65e1ca6f08539e5ed7b278bf25fdf9162670956fe35ca0428223c95cd86"} Dec 05 11:08:34.686437 master-0 kubenswrapper[24928]: I1205 11:08:34.682773 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c468788f-s9m47" event={"ID":"c60d3d04-41e1-4823-b209-18644852fbca","Type":"ContainerStarted","Data":"3951c9fef19e689b006d0afb1ac5f92b7012b03a375532b3fc2179bbc5ba3245"} Dec 05 11:08:34.691386 master-0 kubenswrapper[24928]: I1205 11:08:34.689239 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83110f63-0a22-47c7-a52c-0bab96c66f54","Type":"ContainerStarted","Data":"bf8b0b777d70d7799800f411656b070d857518d4e8f2a51a534d0d9b06a3b85b"} Dec 05 11:08:34.691386 master-0 kubenswrapper[24928]: I1205 11:08:34.689308 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83110f63-0a22-47c7-a52c-0bab96c66f54","Type":"ContainerStarted","Data":"26f2caaf07673e6538b87cacc060da73d3cfe6cd2deeb29158b6b3795a161708"} Dec 05 11:08:34.691386 master-0 kubenswrapper[24928]: I1205 11:08:34.689567 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="83110f63-0a22-47c7-a52c-0bab96c66f54" containerName="nova-metadata-log" containerID="cri-o://26f2caaf07673e6538b87cacc060da73d3cfe6cd2deeb29158b6b3795a161708" gracePeriod=30 Dec 05 11:08:34.691386 master-0 kubenswrapper[24928]: I1205 11:08:34.689693 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="83110f63-0a22-47c7-a52c-0bab96c66f54" containerName="nova-metadata-metadata" containerID="cri-o://bf8b0b777d70d7799800f411656b070d857518d4e8f2a51a534d0d9b06a3b85b" gracePeriod=30 Dec 05 11:08:34.700270 master-0 kubenswrapper[24928]: I1205 11:08:34.700178 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.700155268 podStartE2EDuration="3.700155268s" podCreationTimestamp="2025-12-05 11:08:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:34.699052452 +0000 UTC m=+1274.702246303" watchObservedRunningTime="2025-12-05 11:08:34.700155268 +0000 UTC m=+1274.703349129" Dec 05 11:08:34.710845 master-0 kubenswrapper[24928]: I1205 11:08:34.710763 24928 generic.go:334] "Generic (PLEG): container finished" podID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerID="5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d" exitCode=143 Dec 05 11:08:34.711062 master-0 kubenswrapper[24928]: I1205 11:08:34.710918 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62fbcd8d-b895-4135-bc9a-339f4369e359","Type":"ContainerDied","Data":"5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d"} Dec 05 11:08:34.727050 master-0 kubenswrapper[24928]: I1205 11:08:34.726966 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.726945341 podStartE2EDuration="3.726945341s" podCreationTimestamp="2025-12-05 11:08:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:34.719012488 +0000 UTC m=+1274.722206339" watchObservedRunningTime="2025-12-05 11:08:34.726945341 +0000 UTC m=+1274.730139192" Dec 05 11:08:34.824673 master-0 kubenswrapper[24928]: I1205 11:08:34.824571 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.824551262 podStartE2EDuration="2.824551262s" podCreationTimestamp="2025-12-05 11:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:34.802395631 +0000 UTC m=+1274.805589492" watchObservedRunningTime="2025-12-05 11:08:34.824551262 +0000 UTC m=+1274.827745113" Dec 05 11:08:35.719267 master-0 kubenswrapper[24928]: I1205 11:08:35.718189 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:08:35.728620 master-0 kubenswrapper[24928]: I1205 11:08:35.727608 24928 generic.go:334] "Generic (PLEG): container finished" podID="923a2187-d634-44ba-bb89-bd632b6e2908" containerID="5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab" exitCode=0 Dec 05 11:08:35.728620 master-0 kubenswrapper[24928]: I1205 11:08:35.727680 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:08:35.728620 master-0 kubenswrapper[24928]: I1205 11:08:35.727715 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"923a2187-d634-44ba-bb89-bd632b6e2908","Type":"ContainerDied","Data":"5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab"} Dec 05 11:08:35.728620 master-0 kubenswrapper[24928]: I1205 11:08:35.728206 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"923a2187-d634-44ba-bb89-bd632b6e2908","Type":"ContainerDied","Data":"1883ca908e07a5ad552278d8a5ef796feb8a5eac5cbe069929edbd04b5a61254"} Dec 05 11:08:35.728620 master-0 kubenswrapper[24928]: I1205 11:08:35.728230 24928 scope.go:117] "RemoveContainer" containerID="5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab" Dec 05 11:08:35.734321 master-0 kubenswrapper[24928]: I1205 11:08:35.732810 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c468788f-s9m47" event={"ID":"c60d3d04-41e1-4823-b209-18644852fbca","Type":"ContainerStarted","Data":"1082567249fb3db616680c786166550e849e7bfabeb98ebb62514ad7b65db737"} Dec 05 11:08:35.734321 master-0 kubenswrapper[24928]: I1205 11:08:35.734523 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:35.752595 master-0 kubenswrapper[24928]: I1205 11:08:35.752503 24928 generic.go:334] "Generic (PLEG): container finished" podID="83110f63-0a22-47c7-a52c-0bab96c66f54" containerID="bf8b0b777d70d7799800f411656b070d857518d4e8f2a51a534d0d9b06a3b85b" exitCode=0 Dec 05 11:08:35.752595 master-0 kubenswrapper[24928]: I1205 11:08:35.752562 24928 generic.go:334] "Generic (PLEG): container finished" podID="83110f63-0a22-47c7-a52c-0bab96c66f54" containerID="26f2caaf07673e6538b87cacc060da73d3cfe6cd2deeb29158b6b3795a161708" exitCode=143 Dec 05 11:08:35.752968 master-0 kubenswrapper[24928]: I1205 11:08:35.752897 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83110f63-0a22-47c7-a52c-0bab96c66f54","Type":"ContainerDied","Data":"bf8b0b777d70d7799800f411656b070d857518d4e8f2a51a534d0d9b06a3b85b"} Dec 05 11:08:35.753120 master-0 kubenswrapper[24928]: I1205 11:08:35.753089 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83110f63-0a22-47c7-a52c-0bab96c66f54","Type":"ContainerDied","Data":"26f2caaf07673e6538b87cacc060da73d3cfe6cd2deeb29158b6b3795a161708"} Dec 05 11:08:35.761958 master-0 kubenswrapper[24928]: I1205 11:08:35.761899 24928 scope.go:117] "RemoveContainer" containerID="5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab" Dec 05 11:08:35.762388 master-0 kubenswrapper[24928]: I1205 11:08:35.762328 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-config-data\") pod \"923a2187-d634-44ba-bb89-bd632b6e2908\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " Dec 05 11:08:35.762548 master-0 kubenswrapper[24928]: E1205 11:08:35.762511 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab\": container with ID starting with 5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab not found: ID does not exist" containerID="5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab" Dec 05 11:08:35.762631 master-0 kubenswrapper[24928]: I1205 11:08:35.762562 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab"} err="failed to get container status \"5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab\": rpc error: code = NotFound desc = could not find container \"5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab\": container with ID starting with 5c8c4e7c96c0baa54bbf055d068afb325ed3ca4893373faad184b337fcc7c1ab not found: ID does not exist" Dec 05 11:08:35.762769 master-0 kubenswrapper[24928]: I1205 11:08:35.762721 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6fvg\" (UniqueName: \"kubernetes.io/projected/923a2187-d634-44ba-bb89-bd632b6e2908-kube-api-access-w6fvg\") pod \"923a2187-d634-44ba-bb89-bd632b6e2908\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " Dec 05 11:08:35.762877 master-0 kubenswrapper[24928]: I1205 11:08:35.762853 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-combined-ca-bundle\") pod \"923a2187-d634-44ba-bb89-bd632b6e2908\" (UID: \"923a2187-d634-44ba-bb89-bd632b6e2908\") " Dec 05 11:08:35.805670 master-0 kubenswrapper[24928]: I1205 11:08:35.805169 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/923a2187-d634-44ba-bb89-bd632b6e2908-kube-api-access-w6fvg" (OuterVolumeSpecName: "kube-api-access-w6fvg") pod "923a2187-d634-44ba-bb89-bd632b6e2908" (UID: "923a2187-d634-44ba-bb89-bd632b6e2908"). InnerVolumeSpecName "kube-api-access-w6fvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:35.823007 master-0 kubenswrapper[24928]: I1205 11:08:35.822929 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "923a2187-d634-44ba-bb89-bd632b6e2908" (UID: "923a2187-d634-44ba-bb89-bd632b6e2908"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:35.863787 master-0 kubenswrapper[24928]: I1205 11:08:35.863715 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-config-data" (OuterVolumeSpecName: "config-data") pod "923a2187-d634-44ba-bb89-bd632b6e2908" (UID: "923a2187-d634-44ba-bb89-bd632b6e2908"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:35.878650 master-0 kubenswrapper[24928]: I1205 11:08:35.865493 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c468788f-s9m47" podStartSLOduration=3.8654722 podStartE2EDuration="3.8654722s" podCreationTimestamp="2025-12-05 11:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:35.864357363 +0000 UTC m=+1275.867551224" watchObservedRunningTime="2025-12-05 11:08:35.8654722 +0000 UTC m=+1275.868666051" Dec 05 11:08:35.878650 master-0 kubenswrapper[24928]: I1205 11:08:35.867987 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6fvg\" (UniqueName: \"kubernetes.io/projected/923a2187-d634-44ba-bb89-bd632b6e2908-kube-api-access-w6fvg\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:35.878650 master-0 kubenswrapper[24928]: I1205 11:08:35.868028 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:35.878650 master-0 kubenswrapper[24928]: I1205 11:08:35.868060 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/923a2187-d634-44ba-bb89-bd632b6e2908-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:35.941920 master-0 kubenswrapper[24928]: I1205 11:08:35.941876 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:36.075680 master-0 kubenswrapper[24928]: I1205 11:08:36.073688 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-nova-metadata-tls-certs\") pod \"83110f63-0a22-47c7-a52c-0bab96c66f54\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " Dec 05 11:08:36.075680 master-0 kubenswrapper[24928]: I1205 11:08:36.073944 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-combined-ca-bundle\") pod \"83110f63-0a22-47c7-a52c-0bab96c66f54\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " Dec 05 11:08:36.075680 master-0 kubenswrapper[24928]: I1205 11:08:36.073978 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxb95\" (UniqueName: \"kubernetes.io/projected/83110f63-0a22-47c7-a52c-0bab96c66f54-kube-api-access-zxb95\") pod \"83110f63-0a22-47c7-a52c-0bab96c66f54\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " Dec 05 11:08:36.075680 master-0 kubenswrapper[24928]: I1205 11:08:36.074501 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83110f63-0a22-47c7-a52c-0bab96c66f54-logs\") pod \"83110f63-0a22-47c7-a52c-0bab96c66f54\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " Dec 05 11:08:36.075680 master-0 kubenswrapper[24928]: I1205 11:08:36.074664 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-config-data\") pod \"83110f63-0a22-47c7-a52c-0bab96c66f54\" (UID: \"83110f63-0a22-47c7-a52c-0bab96c66f54\") " Dec 05 11:08:36.091479 master-0 kubenswrapper[24928]: I1205 11:08:36.089390 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83110f63-0a22-47c7-a52c-0bab96c66f54-logs" (OuterVolumeSpecName: "logs") pod "83110f63-0a22-47c7-a52c-0bab96c66f54" (UID: "83110f63-0a22-47c7-a52c-0bab96c66f54"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:36.099030 master-0 kubenswrapper[24928]: I1205 11:08:36.095748 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83110f63-0a22-47c7-a52c-0bab96c66f54-kube-api-access-zxb95" (OuterVolumeSpecName: "kube-api-access-zxb95") pod "83110f63-0a22-47c7-a52c-0bab96c66f54" (UID: "83110f63-0a22-47c7-a52c-0bab96c66f54"). InnerVolumeSpecName "kube-api-access-zxb95". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:36.127101 master-0 kubenswrapper[24928]: I1205 11:08:36.126678 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83110f63-0a22-47c7-a52c-0bab96c66f54" (UID: "83110f63-0a22-47c7-a52c-0bab96c66f54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:36.147810 master-0 kubenswrapper[24928]: I1205 11:08:36.147109 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:08:36.154142 master-0 kubenswrapper[24928]: I1205 11:08:36.154074 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-config-data" (OuterVolumeSpecName: "config-data") pod "83110f63-0a22-47c7-a52c-0bab96c66f54" (UID: "83110f63-0a22-47c7-a52c-0bab96c66f54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:36.177806 master-0 kubenswrapper[24928]: I1205 11:08:36.177735 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:08:36.178609 master-0 kubenswrapper[24928]: I1205 11:08:36.178574 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:36.178609 master-0 kubenswrapper[24928]: I1205 11:08:36.178597 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxb95\" (UniqueName: \"kubernetes.io/projected/83110f63-0a22-47c7-a52c-0bab96c66f54-kube-api-access-zxb95\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:36.178609 master-0 kubenswrapper[24928]: I1205 11:08:36.178609 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83110f63-0a22-47c7-a52c-0bab96c66f54-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:36.178776 master-0 kubenswrapper[24928]: I1205 11:08:36.178618 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:36.190164 master-0 kubenswrapper[24928]: I1205 11:08:36.190099 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "83110f63-0a22-47c7-a52c-0bab96c66f54" (UID: "83110f63-0a22-47c7-a52c-0bab96c66f54"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:36.192957 master-0 kubenswrapper[24928]: I1205 11:08:36.192882 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:08:36.193639 master-0 kubenswrapper[24928]: E1205 11:08:36.193604 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83110f63-0a22-47c7-a52c-0bab96c66f54" containerName="nova-metadata-log" Dec 05 11:08:36.193639 master-0 kubenswrapper[24928]: I1205 11:08:36.193633 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="83110f63-0a22-47c7-a52c-0bab96c66f54" containerName="nova-metadata-log" Dec 05 11:08:36.193738 master-0 kubenswrapper[24928]: E1205 11:08:36.193672 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="923a2187-d634-44ba-bb89-bd632b6e2908" containerName="nova-scheduler-scheduler" Dec 05 11:08:36.193738 master-0 kubenswrapper[24928]: I1205 11:08:36.193681 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="923a2187-d634-44ba-bb89-bd632b6e2908" containerName="nova-scheduler-scheduler" Dec 05 11:08:36.193738 master-0 kubenswrapper[24928]: E1205 11:08:36.193708 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83110f63-0a22-47c7-a52c-0bab96c66f54" containerName="nova-metadata-metadata" Dec 05 11:08:36.193738 master-0 kubenswrapper[24928]: I1205 11:08:36.193717 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="83110f63-0a22-47c7-a52c-0bab96c66f54" containerName="nova-metadata-metadata" Dec 05 11:08:36.194026 master-0 kubenswrapper[24928]: I1205 11:08:36.193994 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="83110f63-0a22-47c7-a52c-0bab96c66f54" containerName="nova-metadata-log" Dec 05 11:08:36.194074 master-0 kubenswrapper[24928]: I1205 11:08:36.194051 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="923a2187-d634-44ba-bb89-bd632b6e2908" containerName="nova-scheduler-scheduler" Dec 05 11:08:36.194105 master-0 kubenswrapper[24928]: I1205 11:08:36.194080 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="83110f63-0a22-47c7-a52c-0bab96c66f54" containerName="nova-metadata-metadata" Dec 05 11:08:36.195109 master-0 kubenswrapper[24928]: I1205 11:08:36.195066 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:08:36.197792 master-0 kubenswrapper[24928]: I1205 11:08:36.197755 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 11:08:36.252742 master-0 kubenswrapper[24928]: I1205 11:08:36.226282 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="923a2187-d634-44ba-bb89-bd632b6e2908" path="/var/lib/kubelet/pods/923a2187-d634-44ba-bb89-bd632b6e2908/volumes" Dec 05 11:08:36.287019 master-0 kubenswrapper[24928]: I1205 11:08:36.281212 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-config-data\") pod \"nova-scheduler-0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:36.287019 master-0 kubenswrapper[24928]: I1205 11:08:36.281276 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:36.287019 master-0 kubenswrapper[24928]: I1205 11:08:36.281653 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76gqm\" (UniqueName: \"kubernetes.io/projected/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-kube-api-access-76gqm\") pod \"nova-scheduler-0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:36.287019 master-0 kubenswrapper[24928]: I1205 11:08:36.281799 24928 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/83110f63-0a22-47c7-a52c-0bab96c66f54-nova-metadata-tls-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:36.287019 master-0 kubenswrapper[24928]: I1205 11:08:36.284526 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:08:36.383726 master-0 kubenswrapper[24928]: I1205 11:08:36.383589 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76gqm\" (UniqueName: \"kubernetes.io/projected/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-kube-api-access-76gqm\") pod \"nova-scheduler-0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:36.383950 master-0 kubenswrapper[24928]: I1205 11:08:36.383805 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-config-data\") pod \"nova-scheduler-0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:36.383950 master-0 kubenswrapper[24928]: I1205 11:08:36.383833 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:36.388951 master-0 kubenswrapper[24928]: I1205 11:08:36.388893 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-config-data\") pod \"nova-scheduler-0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:36.390259 master-0 kubenswrapper[24928]: I1205 11:08:36.390218 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:36.412586 master-0 kubenswrapper[24928]: I1205 11:08:36.412505 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76gqm\" (UniqueName: \"kubernetes.io/projected/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-kube-api-access-76gqm\") pod \"nova-scheduler-0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " pod="openstack/nova-scheduler-0" Dec 05 11:08:36.610982 master-0 kubenswrapper[24928]: I1205 11:08:36.610907 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:08:36.794944 master-0 kubenswrapper[24928]: I1205 11:08:36.793895 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:36.794944 master-0 kubenswrapper[24928]: I1205 11:08:36.794455 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"83110f63-0a22-47c7-a52c-0bab96c66f54","Type":"ContainerDied","Data":"29ee894fc75f2bab4ac66b5620bccc681d7007519b2c16581cd71a0be3888faa"} Dec 05 11:08:36.794944 master-0 kubenswrapper[24928]: I1205 11:08:36.794488 24928 scope.go:117] "RemoveContainer" containerID="bf8b0b777d70d7799800f411656b070d857518d4e8f2a51a534d0d9b06a3b85b" Dec 05 11:08:36.847128 master-0 kubenswrapper[24928]: I1205 11:08:36.846698 24928 scope.go:117] "RemoveContainer" containerID="26f2caaf07673e6538b87cacc060da73d3cfe6cd2deeb29158b6b3795a161708" Dec 05 11:08:36.851488 master-0 kubenswrapper[24928]: I1205 11:08:36.850520 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:36.869895 master-0 kubenswrapper[24928]: I1205 11:08:36.869781 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:36.886395 master-0 kubenswrapper[24928]: I1205 11:08:36.885619 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:36.887906 master-0 kubenswrapper[24928]: I1205 11:08:36.887857 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:36.891369 master-0 kubenswrapper[24928]: I1205 11:08:36.889935 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 11:08:36.891369 master-0 kubenswrapper[24928]: I1205 11:08:36.891300 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 11:08:36.918772 master-0 kubenswrapper[24928]: I1205 11:08:36.914703 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:36.996290 master-0 kubenswrapper[24928]: I1205 11:08:36.996215 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:36.997225 master-0 kubenswrapper[24928]: I1205 11:08:36.997134 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-logs\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:36.997435 master-0 kubenswrapper[24928]: I1205 11:08:36.997391 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:36.997658 master-0 kubenswrapper[24928]: I1205 11:08:36.997634 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpmmp\" (UniqueName: \"kubernetes.io/projected/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-kube-api-access-qpmmp\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:36.997808 master-0 kubenswrapper[24928]: I1205 11:08:36.997780 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-config-data\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:37.005695 master-0 kubenswrapper[24928]: I1205 11:08:37.005657 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:37.100076 master-0 kubenswrapper[24928]: I1205 11:08:37.100023 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:37.102037 master-0 kubenswrapper[24928]: I1205 11:08:37.102014 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-logs\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:37.102636 master-0 kubenswrapper[24928]: I1205 11:08:37.102505 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-logs\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:37.102742 master-0 kubenswrapper[24928]: I1205 11:08:37.102575 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:37.102888 master-0 kubenswrapper[24928]: I1205 11:08:37.102872 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpmmp\" (UniqueName: \"kubernetes.io/projected/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-kube-api-access-qpmmp\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:37.103005 master-0 kubenswrapper[24928]: I1205 11:08:37.102990 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-config-data\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:37.105239 master-0 kubenswrapper[24928]: I1205 11:08:37.105190 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:37.105771 master-0 kubenswrapper[24928]: I1205 11:08:37.105712 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:37.107139 master-0 kubenswrapper[24928]: I1205 11:08:37.107078 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-config-data\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:37.197100 master-0 kubenswrapper[24928]: I1205 11:08:37.196940 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpmmp\" (UniqueName: \"kubernetes.io/projected/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-kube-api-access-qpmmp\") pod \"nova-metadata-0\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " pod="openstack/nova-metadata-0" Dec 05 11:08:37.220825 master-0 kubenswrapper[24928]: I1205 11:08:37.220400 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:37.250649 master-0 kubenswrapper[24928]: I1205 11:08:37.249278 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:08:37.520453 master-0 kubenswrapper[24928]: I1205 11:08:37.518003 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:08:37.640027 master-0 kubenswrapper[24928]: I1205 11:08:37.639975 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-combined-ca-bundle\") pod \"62fbcd8d-b895-4135-bc9a-339f4369e359\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " Dec 05 11:08:37.640299 master-0 kubenswrapper[24928]: I1205 11:08:37.640216 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nhx7\" (UniqueName: \"kubernetes.io/projected/62fbcd8d-b895-4135-bc9a-339f4369e359-kube-api-access-4nhx7\") pod \"62fbcd8d-b895-4135-bc9a-339f4369e359\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " Dec 05 11:08:37.640299 master-0 kubenswrapper[24928]: I1205 11:08:37.640257 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-config-data\") pod \"62fbcd8d-b895-4135-bc9a-339f4369e359\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " Dec 05 11:08:37.640412 master-0 kubenswrapper[24928]: I1205 11:08:37.640338 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62fbcd8d-b895-4135-bc9a-339f4369e359-logs\") pod \"62fbcd8d-b895-4135-bc9a-339f4369e359\" (UID: \"62fbcd8d-b895-4135-bc9a-339f4369e359\") " Dec 05 11:08:37.641499 master-0 kubenswrapper[24928]: I1205 11:08:37.641398 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62fbcd8d-b895-4135-bc9a-339f4369e359-logs" (OuterVolumeSpecName: "logs") pod "62fbcd8d-b895-4135-bc9a-339f4369e359" (UID: "62fbcd8d-b895-4135-bc9a-339f4369e359"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:37.645124 master-0 kubenswrapper[24928]: I1205 11:08:37.645060 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62fbcd8d-b895-4135-bc9a-339f4369e359-kube-api-access-4nhx7" (OuterVolumeSpecName: "kube-api-access-4nhx7") pod "62fbcd8d-b895-4135-bc9a-339f4369e359" (UID: "62fbcd8d-b895-4135-bc9a-339f4369e359"). InnerVolumeSpecName "kube-api-access-4nhx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:37.686444 master-0 kubenswrapper[24928]: I1205 11:08:37.685306 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62fbcd8d-b895-4135-bc9a-339f4369e359" (UID: "62fbcd8d-b895-4135-bc9a-339f4369e359"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:37.687648 master-0 kubenswrapper[24928]: I1205 11:08:37.686856 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-config-data" (OuterVolumeSpecName: "config-data") pod "62fbcd8d-b895-4135-bc9a-339f4369e359" (UID: "62fbcd8d-b895-4135-bc9a-339f4369e359"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:37.743963 master-0 kubenswrapper[24928]: I1205 11:08:37.743862 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:37.743963 master-0 kubenswrapper[24928]: I1205 11:08:37.743943 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nhx7\" (UniqueName: \"kubernetes.io/projected/62fbcd8d-b895-4135-bc9a-339f4369e359-kube-api-access-4nhx7\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:37.743963 master-0 kubenswrapper[24928]: I1205 11:08:37.743959 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62fbcd8d-b895-4135-bc9a-339f4369e359-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:37.743963 master-0 kubenswrapper[24928]: I1205 11:08:37.743972 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62fbcd8d-b895-4135-bc9a-339f4369e359-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:37.806593 master-0 kubenswrapper[24928]: I1205 11:08:37.806512 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"42cf2ec9-0902-4907-9b28-8a20dd22a6f0","Type":"ContainerStarted","Data":"2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d"} Dec 05 11:08:37.806593 master-0 kubenswrapper[24928]: I1205 11:08:37.806596 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"42cf2ec9-0902-4907-9b28-8a20dd22a6f0","Type":"ContainerStarted","Data":"5a38539038736e2eac211e49bac25b70736efafe500c6c2e9d689fc42bf9a468"} Dec 05 11:08:37.810641 master-0 kubenswrapper[24928]: I1205 11:08:37.810602 24928 generic.go:334] "Generic (PLEG): container finished" podID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerID="bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247" exitCode=0 Dec 05 11:08:37.810723 master-0 kubenswrapper[24928]: I1205 11:08:37.810640 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:08:37.810723 master-0 kubenswrapper[24928]: I1205 11:08:37.810682 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62fbcd8d-b895-4135-bc9a-339f4369e359","Type":"ContainerDied","Data":"bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247"} Dec 05 11:08:37.810803 master-0 kubenswrapper[24928]: I1205 11:08:37.810733 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62fbcd8d-b895-4135-bc9a-339f4369e359","Type":"ContainerDied","Data":"0bd75a6069efdd0216a63272838195c490886efb465319fb9739be80cbcd30d9"} Dec 05 11:08:37.810803 master-0 kubenswrapper[24928]: I1205 11:08:37.810751 24928 scope.go:117] "RemoveContainer" containerID="bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247" Dec 05 11:08:37.837021 master-0 kubenswrapper[24928]: I1205 11:08:37.836947 24928 scope.go:117] "RemoveContainer" containerID="5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d" Dec 05 11:08:37.857613 master-0 kubenswrapper[24928]: I1205 11:08:37.857560 24928 scope.go:117] "RemoveContainer" containerID="bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247" Dec 05 11:08:37.858152 master-0 kubenswrapper[24928]: E1205 11:08:37.858044 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247\": container with ID starting with bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247 not found: ID does not exist" containerID="bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247" Dec 05 11:08:37.858212 master-0 kubenswrapper[24928]: I1205 11:08:37.858180 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247"} err="failed to get container status \"bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247\": rpc error: code = NotFound desc = could not find container \"bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247\": container with ID starting with bf05bf5d4d711a18d119913a88576538eb014935d9d17627e5f8053986fa5247 not found: ID does not exist" Dec 05 11:08:37.858212 master-0 kubenswrapper[24928]: I1205 11:08:37.858208 24928 scope.go:117] "RemoveContainer" containerID="5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d" Dec 05 11:08:37.858979 master-0 kubenswrapper[24928]: E1205 11:08:37.858924 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d\": container with ID starting with 5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d not found: ID does not exist" containerID="5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d" Dec 05 11:08:37.859035 master-0 kubenswrapper[24928]: I1205 11:08:37.858977 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d"} err="failed to get container status \"5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d\": rpc error: code = NotFound desc = could not find container \"5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d\": container with ID starting with 5ac70a0c96531df6a0b1cc6744361016e7685649fd5af49db0f03127a00e890d not found: ID does not exist" Dec 05 11:08:38.061977 master-0 kubenswrapper[24928]: I1205 11:08:38.061923 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:38.077039 master-0 kubenswrapper[24928]: I1205 11:08:38.076921 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.076893706 podStartE2EDuration="2.076893706s" podCreationTimestamp="2025-12-05 11:08:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:38.049525608 +0000 UTC m=+1278.052719489" watchObservedRunningTime="2025-12-05 11:08:38.076893706 +0000 UTC m=+1278.080087587" Dec 05 11:08:38.219282 master-0 kubenswrapper[24928]: I1205 11:08:38.219225 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83110f63-0a22-47c7-a52c-0bab96c66f54" path="/var/lib/kubelet/pods/83110f63-0a22-47c7-a52c-0bab96c66f54/volumes" Dec 05 11:08:38.220163 master-0 kubenswrapper[24928]: I1205 11:08:38.220135 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:38.223305 master-0 kubenswrapper[24928]: I1205 11:08:38.223219 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:38.583645 master-0 kubenswrapper[24928]: I1205 11:08:38.583590 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:38.584100 master-0 kubenswrapper[24928]: E1205 11:08:38.584059 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerName="nova-api-api" Dec 05 11:08:38.584100 master-0 kubenswrapper[24928]: I1205 11:08:38.584093 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerName="nova-api-api" Dec 05 11:08:38.584189 master-0 kubenswrapper[24928]: E1205 11:08:38.584111 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerName="nova-api-log" Dec 05 11:08:38.584189 master-0 kubenswrapper[24928]: I1205 11:08:38.584118 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerName="nova-api-log" Dec 05 11:08:38.584398 master-0 kubenswrapper[24928]: I1205 11:08:38.584374 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerName="nova-api-api" Dec 05 11:08:38.584398 master-0 kubenswrapper[24928]: I1205 11:08:38.584394 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="62fbcd8d-b895-4135-bc9a-339f4369e359" containerName="nova-api-log" Dec 05 11:08:38.585604 master-0 kubenswrapper[24928]: I1205 11:08:38.585578 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:08:38.588093 master-0 kubenswrapper[24928]: I1205 11:08:38.588033 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 11:08:38.588275 master-0 kubenswrapper[24928]: I1205 11:08:38.588257 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 05 11:08:38.588462 master-0 kubenswrapper[24928]: I1205 11:08:38.588441 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 05 11:08:38.679243 master-0 kubenswrapper[24928]: I1205 11:08:38.679191 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-public-tls-certs\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.679551 master-0 kubenswrapper[24928]: I1205 11:08:38.679303 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.679551 master-0 kubenswrapper[24928]: I1205 11:08:38.679324 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589956f3-e2cc-4f05-9fd3-1088680bdd7b-logs\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.679551 master-0 kubenswrapper[24928]: I1205 11:08:38.679341 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nqs2\" (UniqueName: \"kubernetes.io/projected/589956f3-e2cc-4f05-9fd3-1088680bdd7b-kube-api-access-8nqs2\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.679551 master-0 kubenswrapper[24928]: I1205 11:08:38.679406 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.679551 master-0 kubenswrapper[24928]: I1205 11:08:38.679470 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-config-data\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.715746 master-0 kubenswrapper[24928]: I1205 11:08:38.715672 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:38.781462 master-0 kubenswrapper[24928]: I1205 11:08:38.781344 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.781547 master-0 kubenswrapper[24928]: I1205 11:08:38.781481 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589956f3-e2cc-4f05-9fd3-1088680bdd7b-logs\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.781547 master-0 kubenswrapper[24928]: I1205 11:08:38.781513 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nqs2\" (UniqueName: \"kubernetes.io/projected/589956f3-e2cc-4f05-9fd3-1088680bdd7b-kube-api-access-8nqs2\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.781692 master-0 kubenswrapper[24928]: I1205 11:08:38.781663 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.781748 master-0 kubenswrapper[24928]: I1205 11:08:38.781713 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-config-data\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.781805 master-0 kubenswrapper[24928]: I1205 11:08:38.781796 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-public-tls-certs\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.786294 master-0 kubenswrapper[24928]: I1205 11:08:38.786237 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589956f3-e2cc-4f05-9fd3-1088680bdd7b-logs\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.786294 master-0 kubenswrapper[24928]: I1205 11:08:38.786279 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.790199 master-0 kubenswrapper[24928]: I1205 11:08:38.790109 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-public-tls-certs\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.799163 master-0 kubenswrapper[24928]: I1205 11:08:38.798785 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-config-data\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.799705 master-0 kubenswrapper[24928]: I1205 11:08:38.799499 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.826749 master-0 kubenswrapper[24928]: I1205 11:08:38.826577 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ba9370a-69d2-4aed-8634-3e2cccb40a1a","Type":"ContainerStarted","Data":"1032239f991b5dd64863b2da71aa9ff464cfd38ec4dd44c0fc941f828f1861c7"} Dec 05 11:08:38.826749 master-0 kubenswrapper[24928]: I1205 11:08:38.826644 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ba9370a-69d2-4aed-8634-3e2cccb40a1a","Type":"ContainerStarted","Data":"bd9b664df541e96920405d11cee963b9beff4aef336d4bf6e3eda1765d7ddbf4"} Dec 05 11:08:38.844542 master-0 kubenswrapper[24928]: I1205 11:08:38.844443 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nqs2\" (UniqueName: \"kubernetes.io/projected/589956f3-e2cc-4f05-9fd3-1088680bdd7b-kube-api-access-8nqs2\") pod \"nova-api-0\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " pod="openstack/nova-api-0" Dec 05 11:08:38.905857 master-0 kubenswrapper[24928]: I1205 11:08:38.905701 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:08:39.434405 master-0 kubenswrapper[24928]: I1205 11:08:39.433937 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:39.838939 master-0 kubenswrapper[24928]: I1205 11:08:39.838880 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589956f3-e2cc-4f05-9fd3-1088680bdd7b","Type":"ContainerStarted","Data":"9100d422acb02e3be2478df3820b7e706e42eaf5df94ba61f88e701d46253663"} Dec 05 11:08:39.838939 master-0 kubenswrapper[24928]: I1205 11:08:39.838932 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589956f3-e2cc-4f05-9fd3-1088680bdd7b","Type":"ContainerStarted","Data":"fa581f2ac92ac11bf6e5da980205539a79fed6725d464b24838c421a9c745989"} Dec 05 11:08:39.840977 master-0 kubenswrapper[24928]: I1205 11:08:39.840925 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ba9370a-69d2-4aed-8634-3e2cccb40a1a","Type":"ContainerStarted","Data":"392291a133ebc27c8d8ddc4a3d83bacedb159f8eb785eea8d6ddfd914eae224e"} Dec 05 11:08:39.875503 master-0 kubenswrapper[24928]: I1205 11:08:39.875402 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.87538171 podStartE2EDuration="3.87538171s" podCreationTimestamp="2025-12-05 11:08:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:39.87290902 +0000 UTC m=+1279.876102871" watchObservedRunningTime="2025-12-05 11:08:39.87538171 +0000 UTC m=+1279.878575561" Dec 05 11:08:40.230609 master-0 kubenswrapper[24928]: I1205 11:08:40.230225 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62fbcd8d-b895-4135-bc9a-339f4369e359" path="/var/lib/kubelet/pods/62fbcd8d-b895-4135-bc9a-339f4369e359/volumes" Dec 05 11:08:40.868555 master-0 kubenswrapper[24928]: I1205 11:08:40.868452 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589956f3-e2cc-4f05-9fd3-1088680bdd7b","Type":"ContainerStarted","Data":"8aab248a4a18b20f40b672910717a747856c3313e10a8afffd34ba7affb7fc60"} Dec 05 11:08:40.898549 master-0 kubenswrapper[24928]: I1205 11:08:40.898077 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.8980564429999998 podStartE2EDuration="2.898056443s" podCreationTimestamp="2025-12-05 11:08:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:40.888902649 +0000 UTC m=+1280.892096540" watchObservedRunningTime="2025-12-05 11:08:40.898056443 +0000 UTC m=+1280.901250294" Dec 05 11:08:41.611292 master-0 kubenswrapper[24928]: I1205 11:08:41.611091 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 11:08:42.005276 master-0 kubenswrapper[24928]: I1205 11:08:42.005133 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:42.029916 master-0 kubenswrapper[24928]: I1205 11:08:42.029847 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:42.221523 master-0 kubenswrapper[24928]: I1205 11:08:42.221411 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 11:08:42.221523 master-0 kubenswrapper[24928]: I1205 11:08:42.221492 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 11:08:42.916389 master-0 kubenswrapper[24928]: I1205 11:08:42.914659 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 05 11:08:42.939547 master-0 kubenswrapper[24928]: I1205 11:08:42.939408 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 05 11:08:43.296860 master-0 kubenswrapper[24928]: I1205 11:08:43.296795 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c468788f-s9m47" Dec 05 11:08:43.617542 master-0 kubenswrapper[24928]: I1205 11:08:43.614563 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76dc4d7685-49kgr"] Dec 05 11:08:43.617542 master-0 kubenswrapper[24928]: I1205 11:08:43.614869 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" podUID="9e744111-e3f8-4e33-8fc5-eb86882de858" containerName="dnsmasq-dns" containerID="cri-o://257c7d15b34cdf394b87a4c3e5275102b5afe40f2ad9d00c6788a5b1197ec38f" gracePeriod=10 Dec 05 11:08:43.919672 master-0 kubenswrapper[24928]: I1205 11:08:43.919582 24928 generic.go:334] "Generic (PLEG): container finished" podID="9e744111-e3f8-4e33-8fc5-eb86882de858" containerID="257c7d15b34cdf394b87a4c3e5275102b5afe40f2ad9d00c6788a5b1197ec38f" exitCode=0 Dec 05 11:08:43.919922 master-0 kubenswrapper[24928]: I1205 11:08:43.919720 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" event={"ID":"9e744111-e3f8-4e33-8fc5-eb86882de858","Type":"ContainerDied","Data":"257c7d15b34cdf394b87a4c3e5275102b5afe40f2ad9d00c6788a5b1197ec38f"} Dec 05 11:08:44.343452 master-0 kubenswrapper[24928]: I1205 11:08:44.343367 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:44.472971 master-0 kubenswrapper[24928]: I1205 11:08:44.472611 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-config\") pod \"9e744111-e3f8-4e33-8fc5-eb86882de858\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " Dec 05 11:08:44.472971 master-0 kubenswrapper[24928]: I1205 11:08:44.472748 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-svc\") pod \"9e744111-e3f8-4e33-8fc5-eb86882de858\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " Dec 05 11:08:44.472971 master-0 kubenswrapper[24928]: I1205 11:08:44.472828 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-sb\") pod \"9e744111-e3f8-4e33-8fc5-eb86882de858\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " Dec 05 11:08:44.472971 master-0 kubenswrapper[24928]: I1205 11:08:44.472928 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-nb\") pod \"9e744111-e3f8-4e33-8fc5-eb86882de858\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " Dec 05 11:08:44.472971 master-0 kubenswrapper[24928]: I1205 11:08:44.472973 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmq2h\" (UniqueName: \"kubernetes.io/projected/9e744111-e3f8-4e33-8fc5-eb86882de858-kube-api-access-xmq2h\") pod \"9e744111-e3f8-4e33-8fc5-eb86882de858\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " Dec 05 11:08:44.473386 master-0 kubenswrapper[24928]: I1205 11:08:44.473001 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-swift-storage-0\") pod \"9e744111-e3f8-4e33-8fc5-eb86882de858\" (UID: \"9e744111-e3f8-4e33-8fc5-eb86882de858\") " Dec 05 11:08:44.484477 master-0 kubenswrapper[24928]: I1205 11:08:44.484394 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e744111-e3f8-4e33-8fc5-eb86882de858-kube-api-access-xmq2h" (OuterVolumeSpecName: "kube-api-access-xmq2h") pod "9e744111-e3f8-4e33-8fc5-eb86882de858" (UID: "9e744111-e3f8-4e33-8fc5-eb86882de858"). InnerVolumeSpecName "kube-api-access-xmq2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:44.601447 master-0 kubenswrapper[24928]: I1205 11:08:44.584413 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmq2h\" (UniqueName: \"kubernetes.io/projected/9e744111-e3f8-4e33-8fc5-eb86882de858-kube-api-access-xmq2h\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:44.615032 master-0 kubenswrapper[24928]: I1205 11:08:44.612574 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e744111-e3f8-4e33-8fc5-eb86882de858" (UID: "9e744111-e3f8-4e33-8fc5-eb86882de858"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44.646887 master-0 kubenswrapper[24928]: I1205 11:08:44.645136 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9e744111-e3f8-4e33-8fc5-eb86882de858" (UID: "9e744111-e3f8-4e33-8fc5-eb86882de858"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44.650445 master-0 kubenswrapper[24928]: I1205 11:08:44.649055 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9e744111-e3f8-4e33-8fc5-eb86882de858" (UID: "9e744111-e3f8-4e33-8fc5-eb86882de858"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44.696446 master-0 kubenswrapper[24928]: I1205 11:08:44.686862 24928 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-svc\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:44.696446 master-0 kubenswrapper[24928]: I1205 11:08:44.686922 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-sb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:44.696446 master-0 kubenswrapper[24928]: I1205 11:08:44.686938 24928 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-dns-swift-storage-0\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:44.708297 master-0 kubenswrapper[24928]: I1205 11:08:44.697541 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-v6rjc"] Dec 05 11:08:44.708297 master-0 kubenswrapper[24928]: E1205 11:08:44.698113 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e744111-e3f8-4e33-8fc5-eb86882de858" containerName="init" Dec 05 11:08:44.708297 master-0 kubenswrapper[24928]: I1205 11:08:44.698128 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e744111-e3f8-4e33-8fc5-eb86882de858" containerName="init" Dec 05 11:08:44.708297 master-0 kubenswrapper[24928]: E1205 11:08:44.698197 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e744111-e3f8-4e33-8fc5-eb86882de858" containerName="dnsmasq-dns" Dec 05 11:08:44.708297 master-0 kubenswrapper[24928]: I1205 11:08:44.698207 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e744111-e3f8-4e33-8fc5-eb86882de858" containerName="dnsmasq-dns" Dec 05 11:08:44.708297 master-0 kubenswrapper[24928]: I1205 11:08:44.698475 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e744111-e3f8-4e33-8fc5-eb86882de858" containerName="dnsmasq-dns" Dec 05 11:08:44.708297 master-0 kubenswrapper[24928]: I1205 11:08:44.699300 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.708790 master-0 kubenswrapper[24928]: I1205 11:08:44.708572 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 05 11:08:44.708790 master-0 kubenswrapper[24928]: I1205 11:08:44.708775 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 05 11:08:44.726580 master-0 kubenswrapper[24928]: I1205 11:08:44.719384 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9e744111-e3f8-4e33-8fc5-eb86882de858" (UID: "9e744111-e3f8-4e33-8fc5-eb86882de858"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44.726580 master-0 kubenswrapper[24928]: I1205 11:08:44.719618 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-host-discover-clfhv"] Dec 05 11:08:44.726580 master-0 kubenswrapper[24928]: I1205 11:08:44.721842 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.751446 master-0 kubenswrapper[24928]: I1205 11:08:44.748387 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-v6rjc"] Dec 05 11:08:44.787174 master-0 kubenswrapper[24928]: I1205 11:08:44.787119 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-host-discover-clfhv"] Dec 05 11:08:44.789317 master-0 kubenswrapper[24928]: I1205 11:08:44.789253 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-scripts\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.789383 master-0 kubenswrapper[24928]: I1205 11:08:44.789327 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.789417 master-0 kubenswrapper[24928]: I1205 11:08:44.789397 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-config-data\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.789672 master-0 kubenswrapper[24928]: I1205 11:08:44.789640 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-combined-ca-bundle\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.789768 master-0 kubenswrapper[24928]: I1205 11:08:44.789734 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwn42\" (UniqueName: \"kubernetes.io/projected/d2f41a31-1432-4cb3-a605-b55952b32508-kube-api-access-lwn42\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.789888 master-0 kubenswrapper[24928]: I1205 11:08:44.789860 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-scripts\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.789930 master-0 kubenswrapper[24928]: I1205 11:08:44.789908 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hzgx\" (UniqueName: \"kubernetes.io/projected/f9837315-d749-43df-b3b0-a7c36eabec7e-kube-api-access-2hzgx\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.789987 master-0 kubenswrapper[24928]: I1205 11:08:44.789968 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-config-data\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.790400 master-0 kubenswrapper[24928]: I1205 11:08:44.790367 24928 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-ovsdbserver-nb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:44.823945 master-0 kubenswrapper[24928]: I1205 11:08:44.823871 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-config" (OuterVolumeSpecName: "config") pod "9e744111-e3f8-4e33-8fc5-eb86882de858" (UID: "9e744111-e3f8-4e33-8fc5-eb86882de858"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:08:44.892683 master-0 kubenswrapper[24928]: I1205 11:08:44.892602 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwn42\" (UniqueName: \"kubernetes.io/projected/d2f41a31-1432-4cb3-a605-b55952b32508-kube-api-access-lwn42\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.892911 master-0 kubenswrapper[24928]: I1205 11:08:44.892738 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-scripts\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.892911 master-0 kubenswrapper[24928]: I1205 11:08:44.892768 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hzgx\" (UniqueName: \"kubernetes.io/projected/f9837315-d749-43df-b3b0-a7c36eabec7e-kube-api-access-2hzgx\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.892911 master-0 kubenswrapper[24928]: I1205 11:08:44.892809 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-config-data\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.892911 master-0 kubenswrapper[24928]: I1205 11:08:44.892846 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-scripts\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.892911 master-0 kubenswrapper[24928]: I1205 11:08:44.892871 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.892911 master-0 kubenswrapper[24928]: I1205 11:08:44.892912 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-config-data\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.893196 master-0 kubenswrapper[24928]: I1205 11:08:44.893036 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-combined-ca-bundle\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.893196 master-0 kubenswrapper[24928]: I1205 11:08:44.893169 24928 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e744111-e3f8-4e33-8fc5-eb86882de858-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:44.897696 master-0 kubenswrapper[24928]: I1205 11:08:44.897633 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-combined-ca-bundle\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.899523 master-0 kubenswrapper[24928]: I1205 11:08:44.899444 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-scripts\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.901224 master-0 kubenswrapper[24928]: I1205 11:08:44.901194 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-config-data\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.901325 master-0 kubenswrapper[24928]: I1205 11:08:44.901222 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-scripts\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.901448 master-0 kubenswrapper[24928]: I1205 11:08:44.901354 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.905661 master-0 kubenswrapper[24928]: I1205 11:08:44.905622 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-config-data\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.909163 master-0 kubenswrapper[24928]: I1205 11:08:44.909114 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwn42\" (UniqueName: \"kubernetes.io/projected/d2f41a31-1432-4cb3-a605-b55952b32508-kube-api-access-lwn42\") pod \"nova-cell1-host-discover-clfhv\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:44.909362 master-0 kubenswrapper[24928]: I1205 11:08:44.909308 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hzgx\" (UniqueName: \"kubernetes.io/projected/f9837315-d749-43df-b3b0-a7c36eabec7e-kube-api-access-2hzgx\") pod \"nova-cell1-cell-mapping-v6rjc\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:44.933988 master-0 kubenswrapper[24928]: I1205 11:08:44.933903 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" event={"ID":"9e744111-e3f8-4e33-8fc5-eb86882de858","Type":"ContainerDied","Data":"e5f82097a45f0ed340361d0c4d462d2694a3662c765b49ba8333e54f96d927d5"} Dec 05 11:08:44.933988 master-0 kubenswrapper[24928]: I1205 11:08:44.933960 24928 scope.go:117] "RemoveContainer" containerID="257c7d15b34cdf394b87a4c3e5275102b5afe40f2ad9d00c6788a5b1197ec38f" Dec 05 11:08:44.934452 master-0 kubenswrapper[24928]: I1205 11:08:44.934045 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76dc4d7685-49kgr" Dec 05 11:08:45.052434 master-0 kubenswrapper[24928]: I1205 11:08:45.052357 24928 scope.go:117] "RemoveContainer" containerID="d20dcecce2abb084fb2107251949a872ca38702bf882d2086a76918214f0de4a" Dec 05 11:08:45.106846 master-0 kubenswrapper[24928]: I1205 11:08:45.106798 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76dc4d7685-49kgr"] Dec 05 11:08:45.152086 master-0 kubenswrapper[24928]: I1205 11:08:45.152010 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76dc4d7685-49kgr"] Dec 05 11:08:45.172993 master-0 kubenswrapper[24928]: I1205 11:08:45.172901 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:45.183401 master-0 kubenswrapper[24928]: I1205 11:08:45.183296 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:45.732397 master-0 kubenswrapper[24928]: W1205 11:08:45.732342 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9837315_d749_43df_b3b0_a7c36eabec7e.slice/crio-a2febad12f9be520d68fd1bed3c0ea31a5248ce1537b8c3c9d39d0b043885a48 WatchSource:0}: Error finding container a2febad12f9be520d68fd1bed3c0ea31a5248ce1537b8c3c9d39d0b043885a48: Status 404 returned error can't find the container with id a2febad12f9be520d68fd1bed3c0ea31a5248ce1537b8c3c9d39d0b043885a48 Dec 05 11:08:45.739337 master-0 kubenswrapper[24928]: I1205 11:08:45.739312 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-v6rjc"] Dec 05 11:08:45.925028 master-0 kubenswrapper[24928]: I1205 11:08:45.924987 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-host-discover-clfhv"] Dec 05 11:08:45.964823 master-0 kubenswrapper[24928]: I1205 11:08:45.964772 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-host-discover-clfhv" event={"ID":"d2f41a31-1432-4cb3-a605-b55952b32508","Type":"ContainerStarted","Data":"073b4c762674d95667dbca67d6c0080018a78cbabc7f783b844bfa40caf1d3c7"} Dec 05 11:08:45.966811 master-0 kubenswrapper[24928]: I1205 11:08:45.966732 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v6rjc" event={"ID":"f9837315-d749-43df-b3b0-a7c36eabec7e","Type":"ContainerStarted","Data":"a2febad12f9be520d68fd1bed3c0ea31a5248ce1537b8c3c9d39d0b043885a48"} Dec 05 11:08:46.227359 master-0 kubenswrapper[24928]: I1205 11:08:46.227243 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e744111-e3f8-4e33-8fc5-eb86882de858" path="/var/lib/kubelet/pods/9e744111-e3f8-4e33-8fc5-eb86882de858/volumes" Dec 05 11:08:46.611718 master-0 kubenswrapper[24928]: I1205 11:08:46.611648 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 11:08:46.651077 master-0 kubenswrapper[24928]: I1205 11:08:46.650985 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 11:08:46.983603 master-0 kubenswrapper[24928]: I1205 11:08:46.983369 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v6rjc" event={"ID":"f9837315-d749-43df-b3b0-a7c36eabec7e","Type":"ContainerStarted","Data":"e646ca28ebe6fb3f9f871e91271ced75c5814530553a28bcb695d05f7558360c"} Dec 05 11:08:46.988623 master-0 kubenswrapper[24928]: I1205 11:08:46.988539 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-host-discover-clfhv" event={"ID":"d2f41a31-1432-4cb3-a605-b55952b32508","Type":"ContainerStarted","Data":"19d0eb0603192eef0552d5244e0df029fbe0013c6cdb400cfaa38b417c70137a"} Dec 05 11:08:47.022468 master-0 kubenswrapper[24928]: I1205 11:08:47.021402 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-v6rjc" podStartSLOduration=3.021376759 podStartE2EDuration="3.021376759s" podCreationTimestamp="2025-12-05 11:08:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:47.009802857 +0000 UTC m=+1287.012996738" watchObservedRunningTime="2025-12-05 11:08:47.021376759 +0000 UTC m=+1287.024570610" Dec 05 11:08:47.031348 master-0 kubenswrapper[24928]: I1205 11:08:47.031263 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 11:08:47.042786 master-0 kubenswrapper[24928]: I1205 11:08:47.042669 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-host-discover-clfhv" podStartSLOduration=3.042641678 podStartE2EDuration="3.042641678s" podCreationTimestamp="2025-12-05 11:08:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:08:47.033248559 +0000 UTC m=+1287.036442430" watchObservedRunningTime="2025-12-05 11:08:47.042641678 +0000 UTC m=+1287.045835529" Dec 05 11:08:47.222517 master-0 kubenswrapper[24928]: I1205 11:08:47.222380 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 11:08:47.222517 master-0 kubenswrapper[24928]: I1205 11:08:47.222492 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 11:08:48.238807 master-0 kubenswrapper[24928]: I1205 11:08:48.238713 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.128.1.9:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:08:48.239403 master-0 kubenswrapper[24928]: I1205 11:08:48.238834 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.128.1.9:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:08:48.909560 master-0 kubenswrapper[24928]: I1205 11:08:48.909488 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:08:48.909801 master-0 kubenswrapper[24928]: I1205 11:08:48.909607 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:08:49.024791 master-0 kubenswrapper[24928]: I1205 11:08:49.024730 24928 generic.go:334] "Generic (PLEG): container finished" podID="d2f41a31-1432-4cb3-a605-b55952b32508" containerID="19d0eb0603192eef0552d5244e0df029fbe0013c6cdb400cfaa38b417c70137a" exitCode=0 Dec 05 11:08:49.025019 master-0 kubenswrapper[24928]: I1205 11:08:49.024794 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-host-discover-clfhv" event={"ID":"d2f41a31-1432-4cb3-a605-b55952b32508","Type":"ContainerDied","Data":"19d0eb0603192eef0552d5244e0df029fbe0013c6cdb400cfaa38b417c70137a"} Dec 05 11:08:49.918669 master-0 kubenswrapper[24928]: I1205 11:08:49.918585 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.128.1.10:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:08:49.919383 master-0 kubenswrapper[24928]: I1205 11:08:49.918592 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.128.1.10:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:08:50.589295 master-0 kubenswrapper[24928]: I1205 11:08:50.589226 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:50.701351 master-0 kubenswrapper[24928]: I1205 11:08:50.701141 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwn42\" (UniqueName: \"kubernetes.io/projected/d2f41a31-1432-4cb3-a605-b55952b32508-kube-api-access-lwn42\") pod \"d2f41a31-1432-4cb3-a605-b55952b32508\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " Dec 05 11:08:50.701351 master-0 kubenswrapper[24928]: I1205 11:08:50.701259 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-combined-ca-bundle\") pod \"d2f41a31-1432-4cb3-a605-b55952b32508\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " Dec 05 11:08:50.701351 master-0 kubenswrapper[24928]: I1205 11:08:50.701349 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-scripts\") pod \"d2f41a31-1432-4cb3-a605-b55952b32508\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " Dec 05 11:08:50.701734 master-0 kubenswrapper[24928]: I1205 11:08:50.701472 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-config-data\") pod \"d2f41a31-1432-4cb3-a605-b55952b32508\" (UID: \"d2f41a31-1432-4cb3-a605-b55952b32508\") " Dec 05 11:08:50.709843 master-0 kubenswrapper[24928]: I1205 11:08:50.709762 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2f41a31-1432-4cb3-a605-b55952b32508-kube-api-access-lwn42" (OuterVolumeSpecName: "kube-api-access-lwn42") pod "d2f41a31-1432-4cb3-a605-b55952b32508" (UID: "d2f41a31-1432-4cb3-a605-b55952b32508"). InnerVolumeSpecName "kube-api-access-lwn42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:50.710030 master-0 kubenswrapper[24928]: I1205 11:08:50.709807 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-scripts" (OuterVolumeSpecName: "scripts") pod "d2f41a31-1432-4cb3-a605-b55952b32508" (UID: "d2f41a31-1432-4cb3-a605-b55952b32508"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:50.752815 master-0 kubenswrapper[24928]: I1205 11:08:50.752737 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2f41a31-1432-4cb3-a605-b55952b32508" (UID: "d2f41a31-1432-4cb3-a605-b55952b32508"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:50.755909 master-0 kubenswrapper[24928]: I1205 11:08:50.755775 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-config-data" (OuterVolumeSpecName: "config-data") pod "d2f41a31-1432-4cb3-a605-b55952b32508" (UID: "d2f41a31-1432-4cb3-a605-b55952b32508"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:50.805198 master-0 kubenswrapper[24928]: I1205 11:08:50.805107 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:50.805198 master-0 kubenswrapper[24928]: I1205 11:08:50.805168 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:50.805198 master-0 kubenswrapper[24928]: I1205 11:08:50.805181 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwn42\" (UniqueName: \"kubernetes.io/projected/d2f41a31-1432-4cb3-a605-b55952b32508-kube-api-access-lwn42\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:50.805198 master-0 kubenswrapper[24928]: I1205 11:08:50.805190 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2f41a31-1432-4cb3-a605-b55952b32508-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:51.070594 master-0 kubenswrapper[24928]: I1205 11:08:51.070519 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-host-discover-clfhv" event={"ID":"d2f41a31-1432-4cb3-a605-b55952b32508","Type":"ContainerDied","Data":"073b4c762674d95667dbca67d6c0080018a78cbabc7f783b844bfa40caf1d3c7"} Dec 05 11:08:51.070594 master-0 kubenswrapper[24928]: I1205 11:08:51.070576 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="073b4c762674d95667dbca67d6c0080018a78cbabc7f783b844bfa40caf1d3c7" Dec 05 11:08:51.071545 master-0 kubenswrapper[24928]: I1205 11:08:51.070590 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-host-discover-clfhv" Dec 05 11:08:52.084546 master-0 kubenswrapper[24928]: I1205 11:08:52.084406 24928 generic.go:334] "Generic (PLEG): container finished" podID="f9837315-d749-43df-b3b0-a7c36eabec7e" containerID="e646ca28ebe6fb3f9f871e91271ced75c5814530553a28bcb695d05f7558360c" exitCode=0 Dec 05 11:08:52.084546 master-0 kubenswrapper[24928]: I1205 11:08:52.084466 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v6rjc" event={"ID":"f9837315-d749-43df-b3b0-a7c36eabec7e","Type":"ContainerDied","Data":"e646ca28ebe6fb3f9f871e91271ced75c5814530553a28bcb695d05f7558360c"} Dec 05 11:08:53.616821 master-0 kubenswrapper[24928]: I1205 11:08:53.616723 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:53.700169 master-0 kubenswrapper[24928]: I1205 11:08:53.689356 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-scripts\") pod \"f9837315-d749-43df-b3b0-a7c36eabec7e\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " Dec 05 11:08:53.700169 master-0 kubenswrapper[24928]: I1205 11:08:53.689642 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hzgx\" (UniqueName: \"kubernetes.io/projected/f9837315-d749-43df-b3b0-a7c36eabec7e-kube-api-access-2hzgx\") pod \"f9837315-d749-43df-b3b0-a7c36eabec7e\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " Dec 05 11:08:53.700169 master-0 kubenswrapper[24928]: I1205 11:08:53.689764 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-combined-ca-bundle\") pod \"f9837315-d749-43df-b3b0-a7c36eabec7e\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " Dec 05 11:08:53.700169 master-0 kubenswrapper[24928]: I1205 11:08:53.689873 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-config-data\") pod \"f9837315-d749-43df-b3b0-a7c36eabec7e\" (UID: \"f9837315-d749-43df-b3b0-a7c36eabec7e\") " Dec 05 11:08:53.700169 master-0 kubenswrapper[24928]: I1205 11:08:53.693334 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9837315-d749-43df-b3b0-a7c36eabec7e-kube-api-access-2hzgx" (OuterVolumeSpecName: "kube-api-access-2hzgx") pod "f9837315-d749-43df-b3b0-a7c36eabec7e" (UID: "f9837315-d749-43df-b3b0-a7c36eabec7e"). InnerVolumeSpecName "kube-api-access-2hzgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:53.700169 master-0 kubenswrapper[24928]: I1205 11:08:53.693732 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-scripts" (OuterVolumeSpecName: "scripts") pod "f9837315-d749-43df-b3b0-a7c36eabec7e" (UID: "f9837315-d749-43df-b3b0-a7c36eabec7e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:53.722503 master-0 kubenswrapper[24928]: I1205 11:08:53.722309 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9837315-d749-43df-b3b0-a7c36eabec7e" (UID: "f9837315-d749-43df-b3b0-a7c36eabec7e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:53.725532 master-0 kubenswrapper[24928]: I1205 11:08:53.725478 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-config-data" (OuterVolumeSpecName: "config-data") pod "f9837315-d749-43df-b3b0-a7c36eabec7e" (UID: "f9837315-d749-43df-b3b0-a7c36eabec7e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:53.793744 master-0 kubenswrapper[24928]: I1205 11:08:53.793671 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:53.793744 master-0 kubenswrapper[24928]: I1205 11:08:53.793722 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:53.793744 master-0 kubenswrapper[24928]: I1205 11:08:53.793732 24928 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9837315-d749-43df-b3b0-a7c36eabec7e-scripts\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:53.793744 master-0 kubenswrapper[24928]: I1205 11:08:53.793744 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hzgx\" (UniqueName: \"kubernetes.io/projected/f9837315-d749-43df-b3b0-a7c36eabec7e-kube-api-access-2hzgx\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:54.108407 master-0 kubenswrapper[24928]: I1205 11:08:54.108307 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v6rjc" event={"ID":"f9837315-d749-43df-b3b0-a7c36eabec7e","Type":"ContainerDied","Data":"a2febad12f9be520d68fd1bed3c0ea31a5248ce1537b8c3c9d39d0b043885a48"} Dec 05 11:08:54.108407 master-0 kubenswrapper[24928]: I1205 11:08:54.108357 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2febad12f9be520d68fd1bed3c0ea31a5248ce1537b8c3c9d39d0b043885a48" Dec 05 11:08:54.108407 master-0 kubenswrapper[24928]: I1205 11:08:54.108383 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v6rjc" Dec 05 11:08:54.427536 master-0 kubenswrapper[24928]: I1205 11:08:54.426482 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:54.427536 master-0 kubenswrapper[24928]: I1205 11:08:54.426810 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerName="nova-api-log" containerID="cri-o://9100d422acb02e3be2478df3820b7e706e42eaf5df94ba61f88e701d46253663" gracePeriod=30 Dec 05 11:08:54.427536 master-0 kubenswrapper[24928]: I1205 11:08:54.427476 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerName="nova-api-api" containerID="cri-o://8aab248a4a18b20f40b672910717a747856c3313e10a8afffd34ba7affb7fc60" gracePeriod=30 Dec 05 11:08:54.438484 master-0 kubenswrapper[24928]: I1205 11:08:54.438362 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:08:54.438744 master-0 kubenswrapper[24928]: I1205 11:08:54.438679 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="42cf2ec9-0902-4907-9b28-8a20dd22a6f0" containerName="nova-scheduler-scheduler" containerID="cri-o://2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d" gracePeriod=30 Dec 05 11:08:54.449885 master-0 kubenswrapper[24928]: I1205 11:08:54.449759 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:54.450147 master-0 kubenswrapper[24928]: I1205 11:08:54.450042 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerName="nova-metadata-log" containerID="cri-o://1032239f991b5dd64863b2da71aa9ff464cfd38ec4dd44c0fc941f828f1861c7" gracePeriod=30 Dec 05 11:08:54.450217 master-0 kubenswrapper[24928]: I1205 11:08:54.450191 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerName="nova-metadata-metadata" containerID="cri-o://392291a133ebc27c8d8ddc4a3d83bacedb159f8eb785eea8d6ddfd914eae224e" gracePeriod=30 Dec 05 11:08:55.123269 master-0 kubenswrapper[24928]: I1205 11:08:55.123182 24928 generic.go:334] "Generic (PLEG): container finished" podID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerID="9100d422acb02e3be2478df3820b7e706e42eaf5df94ba61f88e701d46253663" exitCode=143 Dec 05 11:08:55.123269 master-0 kubenswrapper[24928]: I1205 11:08:55.123262 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589956f3-e2cc-4f05-9fd3-1088680bdd7b","Type":"ContainerDied","Data":"9100d422acb02e3be2478df3820b7e706e42eaf5df94ba61f88e701d46253663"} Dec 05 11:08:55.127683 master-0 kubenswrapper[24928]: I1205 11:08:55.127616 24928 generic.go:334] "Generic (PLEG): container finished" podID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerID="1032239f991b5dd64863b2da71aa9ff464cfd38ec4dd44c0fc941f828f1861c7" exitCode=143 Dec 05 11:08:55.127828 master-0 kubenswrapper[24928]: I1205 11:08:55.127685 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ba9370a-69d2-4aed-8634-3e2cccb40a1a","Type":"ContainerDied","Data":"1032239f991b5dd64863b2da71aa9ff464cfd38ec4dd44c0fc941f828f1861c7"} Dec 05 11:08:56.613765 master-0 kubenswrapper[24928]: E1205 11:08:56.613691 24928 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:08:56.615884 master-0 kubenswrapper[24928]: E1205 11:08:56.615752 24928 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:08:56.617141 master-0 kubenswrapper[24928]: E1205 11:08:56.617041 24928 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 11:08:56.617204 master-0 kubenswrapper[24928]: E1205 11:08:56.617157 24928 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="42cf2ec9-0902-4907-9b28-8a20dd22a6f0" containerName="nova-scheduler-scheduler" Dec 05 11:08:58.168671 master-0 kubenswrapper[24928]: I1205 11:08:58.168595 24928 generic.go:334] "Generic (PLEG): container finished" podID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerID="8aab248a4a18b20f40b672910717a747856c3313e10a8afffd34ba7affb7fc60" exitCode=0 Dec 05 11:08:58.168671 master-0 kubenswrapper[24928]: I1205 11:08:58.168635 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589956f3-e2cc-4f05-9fd3-1088680bdd7b","Type":"ContainerDied","Data":"8aab248a4a18b20f40b672910717a747856c3313e10a8afffd34ba7affb7fc60"} Dec 05 11:08:58.169219 master-0 kubenswrapper[24928]: I1205 11:08:58.168710 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"589956f3-e2cc-4f05-9fd3-1088680bdd7b","Type":"ContainerDied","Data":"fa581f2ac92ac11bf6e5da980205539a79fed6725d464b24838c421a9c745989"} Dec 05 11:08:58.169219 master-0 kubenswrapper[24928]: I1205 11:08:58.168723 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa581f2ac92ac11bf6e5da980205539a79fed6725d464b24838c421a9c745989" Dec 05 11:08:58.170628 master-0 kubenswrapper[24928]: I1205 11:08:58.170594 24928 generic.go:334] "Generic (PLEG): container finished" podID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerID="392291a133ebc27c8d8ddc4a3d83bacedb159f8eb785eea8d6ddfd914eae224e" exitCode=0 Dec 05 11:08:58.170749 master-0 kubenswrapper[24928]: I1205 11:08:58.170658 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ba9370a-69d2-4aed-8634-3e2cccb40a1a","Type":"ContainerDied","Data":"392291a133ebc27c8d8ddc4a3d83bacedb159f8eb785eea8d6ddfd914eae224e"} Dec 05 11:08:58.246573 master-0 kubenswrapper[24928]: I1205 11:08:58.246370 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:08:58.323635 master-0 kubenswrapper[24928]: I1205 11:08:58.323568 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589956f3-e2cc-4f05-9fd3-1088680bdd7b-logs\") pod \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " Dec 05 11:08:58.324135 master-0 kubenswrapper[24928]: I1205 11:08:58.323677 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nqs2\" (UniqueName: \"kubernetes.io/projected/589956f3-e2cc-4f05-9fd3-1088680bdd7b-kube-api-access-8nqs2\") pod \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " Dec 05 11:08:58.324135 master-0 kubenswrapper[24928]: I1205 11:08:58.323747 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-public-tls-certs\") pod \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " Dec 05 11:08:58.324135 master-0 kubenswrapper[24928]: I1205 11:08:58.323823 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-internal-tls-certs\") pod \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " Dec 05 11:08:58.324135 master-0 kubenswrapper[24928]: I1205 11:08:58.323881 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-config-data\") pod \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " Dec 05 11:08:58.324135 master-0 kubenswrapper[24928]: I1205 11:08:58.323949 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-combined-ca-bundle\") pod \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\" (UID: \"589956f3-e2cc-4f05-9fd3-1088680bdd7b\") " Dec 05 11:08:58.325341 master-0 kubenswrapper[24928]: I1205 11:08:58.325280 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/589956f3-e2cc-4f05-9fd3-1088680bdd7b-logs" (OuterVolumeSpecName: "logs") pod "589956f3-e2cc-4f05-9fd3-1088680bdd7b" (UID: "589956f3-e2cc-4f05-9fd3-1088680bdd7b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:58.328978 master-0 kubenswrapper[24928]: I1205 11:08:58.328854 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/589956f3-e2cc-4f05-9fd3-1088680bdd7b-kube-api-access-8nqs2" (OuterVolumeSpecName: "kube-api-access-8nqs2") pod "589956f3-e2cc-4f05-9fd3-1088680bdd7b" (UID: "589956f3-e2cc-4f05-9fd3-1088680bdd7b"). InnerVolumeSpecName "kube-api-access-8nqs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:58.366236 master-0 kubenswrapper[24928]: I1205 11:08:58.366189 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "589956f3-e2cc-4f05-9fd3-1088680bdd7b" (UID: "589956f3-e2cc-4f05-9fd3-1088680bdd7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:58.374507 master-0 kubenswrapper[24928]: I1205 11:08:58.374458 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-config-data" (OuterVolumeSpecName: "config-data") pod "589956f3-e2cc-4f05-9fd3-1088680bdd7b" (UID: "589956f3-e2cc-4f05-9fd3-1088680bdd7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:58.375560 master-0 kubenswrapper[24928]: I1205 11:08:58.375250 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:58.403878 master-0 kubenswrapper[24928]: I1205 11:08:58.400898 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "589956f3-e2cc-4f05-9fd3-1088680bdd7b" (UID: "589956f3-e2cc-4f05-9fd3-1088680bdd7b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:58.415208 master-0 kubenswrapper[24928]: I1205 11:08:58.414981 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "589956f3-e2cc-4f05-9fd3-1088680bdd7b" (UID: "589956f3-e2cc-4f05-9fd3-1088680bdd7b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:58.427052 master-0 kubenswrapper[24928]: I1205 11:08:58.426955 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-combined-ca-bundle\") pod \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " Dec 05 11:08:58.427309 master-0 kubenswrapper[24928]: I1205 11:08:58.427078 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-nova-metadata-tls-certs\") pod \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " Dec 05 11:08:58.427309 master-0 kubenswrapper[24928]: I1205 11:08:58.427301 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-logs\") pod \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " Dec 05 11:08:58.427408 master-0 kubenswrapper[24928]: I1205 11:08:58.427353 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-config-data\") pod \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " Dec 05 11:08:58.427479 master-0 kubenswrapper[24928]: I1205 11:08:58.427432 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpmmp\" (UniqueName: \"kubernetes.io/projected/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-kube-api-access-qpmmp\") pod \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\" (UID: \"3ba9370a-69d2-4aed-8634-3e2cccb40a1a\") " Dec 05 11:08:58.428250 master-0 kubenswrapper[24928]: I1205 11:08:58.428215 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/589956f3-e2cc-4f05-9fd3-1088680bdd7b-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:58.428250 master-0 kubenswrapper[24928]: I1205 11:08:58.428242 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nqs2\" (UniqueName: \"kubernetes.io/projected/589956f3-e2cc-4f05-9fd3-1088680bdd7b-kube-api-access-8nqs2\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:58.428347 master-0 kubenswrapper[24928]: I1205 11:08:58.428256 24928 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-public-tls-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:58.428347 master-0 kubenswrapper[24928]: I1205 11:08:58.428267 24928 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-internal-tls-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:58.428347 master-0 kubenswrapper[24928]: I1205 11:08:58.428276 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:58.428347 master-0 kubenswrapper[24928]: I1205 11:08:58.428288 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/589956f3-e2cc-4f05-9fd3-1088680bdd7b-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:58.430223 master-0 kubenswrapper[24928]: I1205 11:08:58.430150 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-logs" (OuterVolumeSpecName: "logs") pod "3ba9370a-69d2-4aed-8634-3e2cccb40a1a" (UID: "3ba9370a-69d2-4aed-8634-3e2cccb40a1a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:08:58.432843 master-0 kubenswrapper[24928]: I1205 11:08:58.432765 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-kube-api-access-qpmmp" (OuterVolumeSpecName: "kube-api-access-qpmmp") pod "3ba9370a-69d2-4aed-8634-3e2cccb40a1a" (UID: "3ba9370a-69d2-4aed-8634-3e2cccb40a1a"). InnerVolumeSpecName "kube-api-access-qpmmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:58.474570 master-0 kubenswrapper[24928]: I1205 11:08:58.474473 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-config-data" (OuterVolumeSpecName: "config-data") pod "3ba9370a-69d2-4aed-8634-3e2cccb40a1a" (UID: "3ba9370a-69d2-4aed-8634-3e2cccb40a1a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:58.497523 master-0 kubenswrapper[24928]: I1205 11:08:58.497309 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ba9370a-69d2-4aed-8634-3e2cccb40a1a" (UID: "3ba9370a-69d2-4aed-8634-3e2cccb40a1a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:58.508681 master-0 kubenswrapper[24928]: I1205 11:08:58.508567 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3ba9370a-69d2-4aed-8634-3e2cccb40a1a" (UID: "3ba9370a-69d2-4aed-8634-3e2cccb40a1a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:58.529694 master-0 kubenswrapper[24928]: I1205 11:08:58.529582 24928 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-nova-metadata-tls-certs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:58.529694 master-0 kubenswrapper[24928]: I1205 11:08:58.529632 24928 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-logs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:58.529694 master-0 kubenswrapper[24928]: I1205 11:08:58.529645 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:58.529694 master-0 kubenswrapper[24928]: I1205 11:08:58.529655 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpmmp\" (UniqueName: \"kubernetes.io/projected/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-kube-api-access-qpmmp\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:58.529694 master-0 kubenswrapper[24928]: I1205 11:08:58.529665 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ba9370a-69d2-4aed-8634-3e2cccb40a1a-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:08:59.190567 master-0 kubenswrapper[24928]: I1205 11:08:59.190488 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3ba9370a-69d2-4aed-8634-3e2cccb40a1a","Type":"ContainerDied","Data":"bd9b664df541e96920405d11cee963b9beff4aef336d4bf6e3eda1765d7ddbf4"} Dec 05 11:08:59.191155 master-0 kubenswrapper[24928]: I1205 11:08:59.190578 24928 scope.go:117] "RemoveContainer" containerID="392291a133ebc27c8d8ddc4a3d83bacedb159f8eb785eea8d6ddfd914eae224e" Dec 05 11:08:59.191155 master-0 kubenswrapper[24928]: I1205 11:08:59.190534 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:08:59.191155 master-0 kubenswrapper[24928]: I1205 11:08:59.190534 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:59.235060 master-0 kubenswrapper[24928]: I1205 11:08:59.235011 24928 scope.go:117] "RemoveContainer" containerID="1032239f991b5dd64863b2da71aa9ff464cfd38ec4dd44c0fc941f828f1861c7" Dec 05 11:08:59.245933 master-0 kubenswrapper[24928]: I1205 11:08:59.245821 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:59.275002 master-0 kubenswrapper[24928]: I1205 11:08:59.274851 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:59.297617 master-0 kubenswrapper[24928]: I1205 11:08:59.296855 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:59.322559 master-0 kubenswrapper[24928]: I1205 11:08:59.322345 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:59.324376 master-0 kubenswrapper[24928]: E1205 11:08:59.324320 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerName="nova-api-log" Dec 05 11:08:59.324376 master-0 kubenswrapper[24928]: I1205 11:08:59.324356 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerName="nova-api-log" Dec 05 11:08:59.324376 master-0 kubenswrapper[24928]: E1205 11:08:59.324383 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerName="nova-api-api" Dec 05 11:08:59.324376 master-0 kubenswrapper[24928]: I1205 11:08:59.324390 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerName="nova-api-api" Dec 05 11:08:59.324743 master-0 kubenswrapper[24928]: E1205 11:08:59.324438 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerName="nova-metadata-log" Dec 05 11:08:59.324743 master-0 kubenswrapper[24928]: I1205 11:08:59.324450 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerName="nova-metadata-log" Dec 05 11:08:59.324743 master-0 kubenswrapper[24928]: E1205 11:08:59.324465 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerName="nova-metadata-metadata" Dec 05 11:08:59.324743 master-0 kubenswrapper[24928]: I1205 11:08:59.324472 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerName="nova-metadata-metadata" Dec 05 11:08:59.324743 master-0 kubenswrapper[24928]: E1205 11:08:59.324535 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9837315-d749-43df-b3b0-a7c36eabec7e" containerName="nova-manage" Dec 05 11:08:59.324743 master-0 kubenswrapper[24928]: I1205 11:08:59.324543 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9837315-d749-43df-b3b0-a7c36eabec7e" containerName="nova-manage" Dec 05 11:08:59.324743 master-0 kubenswrapper[24928]: E1205 11:08:59.324570 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2f41a31-1432-4cb3-a605-b55952b32508" containerName="nova-manage" Dec 05 11:08:59.324743 master-0 kubenswrapper[24928]: I1205 11:08:59.324578 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2f41a31-1432-4cb3-a605-b55952b32508" containerName="nova-manage" Dec 05 11:08:59.325074 master-0 kubenswrapper[24928]: I1205 11:08:59.324927 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9837315-d749-43df-b3b0-a7c36eabec7e" containerName="nova-manage" Dec 05 11:08:59.325074 master-0 kubenswrapper[24928]: I1205 11:08:59.324967 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2f41a31-1432-4cb3-a605-b55952b32508" containerName="nova-manage" Dec 05 11:08:59.325074 master-0 kubenswrapper[24928]: I1205 11:08:59.325019 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerName="nova-metadata-log" Dec 05 11:08:59.325074 master-0 kubenswrapper[24928]: I1205 11:08:59.325041 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerName="nova-api-api" Dec 05 11:08:59.325074 master-0 kubenswrapper[24928]: I1205 11:08:59.325072 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" containerName="nova-api-log" Dec 05 11:08:59.325308 master-0 kubenswrapper[24928]: I1205 11:08:59.325086 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" containerName="nova-metadata-metadata" Dec 05 11:08:59.327340 master-0 kubenswrapper[24928]: I1205 11:08:59.327206 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:08:59.331554 master-0 kubenswrapper[24928]: I1205 11:08:59.331468 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 05 11:08:59.338512 master-0 kubenswrapper[24928]: I1205 11:08:59.335219 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 05 11:08:59.338512 master-0 kubenswrapper[24928]: I1205 11:08:59.335571 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 11:08:59.349754 master-0 kubenswrapper[24928]: I1205 11:08:59.347538 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:59.356543 master-0 kubenswrapper[24928]: I1205 11:08:59.356454 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.356543 master-0 kubenswrapper[24928]: I1205 11:08:59.356547 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-internal-tls-certs\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.356974 master-0 kubenswrapper[24928]: I1205 11:08:59.356589 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-config-data\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.356974 master-0 kubenswrapper[24928]: I1205 11:08:59.356669 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-logs\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.356974 master-0 kubenswrapper[24928]: I1205 11:08:59.356806 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdzqc\" (UniqueName: \"kubernetes.io/projected/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-kube-api-access-hdzqc\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.356974 master-0 kubenswrapper[24928]: I1205 11:08:59.356879 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-public-tls-certs\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.380796 master-0 kubenswrapper[24928]: I1205 11:08:59.380584 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:08:59.402311 master-0 kubenswrapper[24928]: I1205 11:08:59.402238 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:59.409345 master-0 kubenswrapper[24928]: I1205 11:08:59.409270 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:08:59.411871 master-0 kubenswrapper[24928]: I1205 11:08:59.411819 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 11:08:59.412744 master-0 kubenswrapper[24928]: I1205 11:08:59.412680 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 11:08:59.421246 master-0 kubenswrapper[24928]: I1205 11:08:59.421141 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.459958 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/610b724c-d8d9-49e5-9760-cb796b37f0e9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.460042 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610b724c-d8d9-49e5-9760-cb796b37f0e9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.460155 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-logs\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.460200 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k95rx\" (UniqueName: \"kubernetes.io/projected/610b724c-d8d9-49e5-9760-cb796b37f0e9-kube-api-access-k95rx\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.460401 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdzqc\" (UniqueName: \"kubernetes.io/projected/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-kube-api-access-hdzqc\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.460460 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610b724c-d8d9-49e5-9760-cb796b37f0e9-config-data\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.460646 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-public-tls-certs\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.461060 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.461703 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-internal-tls-certs\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.461759 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-config-data\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.461794 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/610b724c-d8d9-49e5-9760-cb796b37f0e9-logs\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.465889 master-0 kubenswrapper[24928]: I1205 11:08:59.463247 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-logs\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.466335 master-0 kubenswrapper[24928]: I1205 11:08:59.466289 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-public-tls-certs\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.470720 master-0 kubenswrapper[24928]: I1205 11:08:59.470295 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-internal-tls-certs\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.481783 master-0 kubenswrapper[24928]: I1205 11:08:59.481724 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.483406 master-0 kubenswrapper[24928]: I1205 11:08:59.483334 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdzqc\" (UniqueName: \"kubernetes.io/projected/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-kube-api-access-hdzqc\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.484777 master-0 kubenswrapper[24928]: I1205 11:08:59.484731 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c3336b-a0d7-4e81-b8b2-02bef9bd5984-config-data\") pod \"nova-api-0\" (UID: \"62c3336b-a0d7-4e81-b8b2-02bef9bd5984\") " pod="openstack/nova-api-0" Dec 05 11:08:59.563860 master-0 kubenswrapper[24928]: I1205 11:08:59.563762 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/610b724c-d8d9-49e5-9760-cb796b37f0e9-logs\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.564230 master-0 kubenswrapper[24928]: I1205 11:08:59.564186 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/610b724c-d8d9-49e5-9760-cb796b37f0e9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.564404 master-0 kubenswrapper[24928]: I1205 11:08:59.564386 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610b724c-d8d9-49e5-9760-cb796b37f0e9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.565226 master-0 kubenswrapper[24928]: I1205 11:08:59.564471 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/610b724c-d8d9-49e5-9760-cb796b37f0e9-logs\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.565371 master-0 kubenswrapper[24928]: I1205 11:08:59.565354 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k95rx\" (UniqueName: \"kubernetes.io/projected/610b724c-d8d9-49e5-9760-cb796b37f0e9-kube-api-access-k95rx\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.565765 master-0 kubenswrapper[24928]: I1205 11:08:59.565744 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610b724c-d8d9-49e5-9760-cb796b37f0e9-config-data\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.567918 master-0 kubenswrapper[24928]: I1205 11:08:59.567850 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/610b724c-d8d9-49e5-9760-cb796b37f0e9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.569933 master-0 kubenswrapper[24928]: I1205 11:08:59.569813 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/610b724c-d8d9-49e5-9760-cb796b37f0e9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.570860 master-0 kubenswrapper[24928]: I1205 11:08:59.570836 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/610b724c-d8d9-49e5-9760-cb796b37f0e9-config-data\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:08:59.649345 master-0 kubenswrapper[24928]: I1205 11:08:59.649125 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 11:08:59.821218 master-0 kubenswrapper[24928]: I1205 11:08:59.821148 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:08:59.874450 master-0 kubenswrapper[24928]: I1205 11:08:59.872655 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-76gqm\" (UniqueName: \"kubernetes.io/projected/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-kube-api-access-76gqm\") pod \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " Dec 05 11:08:59.874450 master-0 kubenswrapper[24928]: I1205 11:08:59.872840 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-combined-ca-bundle\") pod \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " Dec 05 11:08:59.874450 master-0 kubenswrapper[24928]: I1205 11:08:59.873090 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-config-data\") pod \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\" (UID: \"42cf2ec9-0902-4907-9b28-8a20dd22a6f0\") " Dec 05 11:08:59.877075 master-0 kubenswrapper[24928]: I1205 11:08:59.876954 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-kube-api-access-76gqm" (OuterVolumeSpecName: "kube-api-access-76gqm") pod "42cf2ec9-0902-4907-9b28-8a20dd22a6f0" (UID: "42cf2ec9-0902-4907-9b28-8a20dd22a6f0"). InnerVolumeSpecName "kube-api-access-76gqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:08:59.907848 master-0 kubenswrapper[24928]: I1205 11:08:59.907696 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42cf2ec9-0902-4907-9b28-8a20dd22a6f0" (UID: "42cf2ec9-0902-4907-9b28-8a20dd22a6f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:08:59.908749 master-0 kubenswrapper[24928]: I1205 11:08:59.908681 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-config-data" (OuterVolumeSpecName: "config-data") pod "42cf2ec9-0902-4907-9b28-8a20dd22a6f0" (UID: "42cf2ec9-0902-4907-9b28-8a20dd22a6f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:09:00.018461 master-0 kubenswrapper[24928]: I1205 11:09:00.004849 24928 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-combined-ca-bundle\") on node \"master-0\" DevicePath \"\"" Dec 05 11:09:00.018461 master-0 kubenswrapper[24928]: I1205 11:09:00.004922 24928 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-config-data\") on node \"master-0\" DevicePath \"\"" Dec 05 11:09:00.018461 master-0 kubenswrapper[24928]: I1205 11:09:00.004936 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-76gqm\" (UniqueName: \"kubernetes.io/projected/42cf2ec9-0902-4907-9b28-8a20dd22a6f0-kube-api-access-76gqm\") on node \"master-0\" DevicePath \"\"" Dec 05 11:09:00.207705 master-0 kubenswrapper[24928]: I1205 11:09:00.207405 24928 generic.go:334] "Generic (PLEG): container finished" podID="42cf2ec9-0902-4907-9b28-8a20dd22a6f0" containerID="2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d" exitCode=0 Dec 05 11:09:00.207705 master-0 kubenswrapper[24928]: I1205 11:09:00.207586 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:09:00.226840 master-0 kubenswrapper[24928]: I1205 11:09:00.226040 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ba9370a-69d2-4aed-8634-3e2cccb40a1a" path="/var/lib/kubelet/pods/3ba9370a-69d2-4aed-8634-3e2cccb40a1a/volumes" Dec 05 11:09:00.227100 master-0 kubenswrapper[24928]: I1205 11:09:00.227063 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="589956f3-e2cc-4f05-9fd3-1088680bdd7b" path="/var/lib/kubelet/pods/589956f3-e2cc-4f05-9fd3-1088680bdd7b/volumes" Dec 05 11:09:00.228589 master-0 kubenswrapper[24928]: I1205 11:09:00.228489 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"42cf2ec9-0902-4907-9b28-8a20dd22a6f0","Type":"ContainerDied","Data":"2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d"} Dec 05 11:09:00.228693 master-0 kubenswrapper[24928]: I1205 11:09:00.228632 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"42cf2ec9-0902-4907-9b28-8a20dd22a6f0","Type":"ContainerDied","Data":"5a38539038736e2eac211e49bac25b70736efafe500c6c2e9d689fc42bf9a468"} Dec 05 11:09:00.228693 master-0 kubenswrapper[24928]: I1205 11:09:00.228676 24928 scope.go:117] "RemoveContainer" containerID="2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d" Dec 05 11:09:00.259257 master-0 kubenswrapper[24928]: I1205 11:09:00.259178 24928 scope.go:117] "RemoveContainer" containerID="2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d" Dec 05 11:09:00.260190 master-0 kubenswrapper[24928]: E1205 11:09:00.260114 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d\": container with ID starting with 2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d not found: ID does not exist" containerID="2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d" Dec 05 11:09:00.260936 master-0 kubenswrapper[24928]: I1205 11:09:00.260185 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d"} err="failed to get container status \"2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d\": rpc error: code = NotFound desc = could not find container \"2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d\": container with ID starting with 2e9374ed7f4ffe6a6e3e690c44173c18069a4eade5fe6a3b77c7d502a2b58b6d not found: ID does not exist" Dec 05 11:09:00.263828 master-0 kubenswrapper[24928]: I1205 11:09:00.263786 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k95rx\" (UniqueName: \"kubernetes.io/projected/610b724c-d8d9-49e5-9760-cb796b37f0e9-kube-api-access-k95rx\") pod \"nova-metadata-0\" (UID: \"610b724c-d8d9-49e5-9760-cb796b37f0e9\") " pod="openstack/nova-metadata-0" Dec 05 11:09:00.363928 master-0 kubenswrapper[24928]: I1205 11:09:00.332866 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 11:09:00.363928 master-0 kubenswrapper[24928]: I1205 11:09:00.346973 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:09:00.363928 master-0 kubenswrapper[24928]: I1205 11:09:00.363538 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:09:00.395524 master-0 kubenswrapper[24928]: I1205 11:09:00.381188 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:09:00.395524 master-0 kubenswrapper[24928]: E1205 11:09:00.381919 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42cf2ec9-0902-4907-9b28-8a20dd22a6f0" containerName="nova-scheduler-scheduler" Dec 05 11:09:00.395524 master-0 kubenswrapper[24928]: I1205 11:09:00.381941 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="42cf2ec9-0902-4907-9b28-8a20dd22a6f0" containerName="nova-scheduler-scheduler" Dec 05 11:09:00.395524 master-0 kubenswrapper[24928]: I1205 11:09:00.382273 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="42cf2ec9-0902-4907-9b28-8a20dd22a6f0" containerName="nova-scheduler-scheduler" Dec 05 11:09:00.395524 master-0 kubenswrapper[24928]: I1205 11:09:00.383328 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:09:00.395524 master-0 kubenswrapper[24928]: I1205 11:09:00.386474 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 11:09:00.396003 master-0 kubenswrapper[24928]: I1205 11:09:00.395562 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:09:00.520518 master-0 kubenswrapper[24928]: I1205 11:09:00.520352 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h85lr\" (UniqueName: \"kubernetes.io/projected/6c1fc53e-ad8f-4b72-b1cd-990241b08a43-kube-api-access-h85lr\") pod \"nova-scheduler-0\" (UID: \"6c1fc53e-ad8f-4b72-b1cd-990241b08a43\") " pod="openstack/nova-scheduler-0" Dec 05 11:09:00.520728 master-0 kubenswrapper[24928]: I1205 11:09:00.520619 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c1fc53e-ad8f-4b72-b1cd-990241b08a43-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6c1fc53e-ad8f-4b72-b1cd-990241b08a43\") " pod="openstack/nova-scheduler-0" Dec 05 11:09:00.521045 master-0 kubenswrapper[24928]: I1205 11:09:00.520991 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c1fc53e-ad8f-4b72-b1cd-990241b08a43-config-data\") pod \"nova-scheduler-0\" (UID: \"6c1fc53e-ad8f-4b72-b1cd-990241b08a43\") " pod="openstack/nova-scheduler-0" Dec 05 11:09:00.561034 master-0 kubenswrapper[24928]: I1205 11:09:00.560970 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 11:09:00.623661 master-0 kubenswrapper[24928]: I1205 11:09:00.623563 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h85lr\" (UniqueName: \"kubernetes.io/projected/6c1fc53e-ad8f-4b72-b1cd-990241b08a43-kube-api-access-h85lr\") pod \"nova-scheduler-0\" (UID: \"6c1fc53e-ad8f-4b72-b1cd-990241b08a43\") " pod="openstack/nova-scheduler-0" Dec 05 11:09:00.623924 master-0 kubenswrapper[24928]: I1205 11:09:00.623713 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c1fc53e-ad8f-4b72-b1cd-990241b08a43-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6c1fc53e-ad8f-4b72-b1cd-990241b08a43\") " pod="openstack/nova-scheduler-0" Dec 05 11:09:00.623924 master-0 kubenswrapper[24928]: I1205 11:09:00.623881 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c1fc53e-ad8f-4b72-b1cd-990241b08a43-config-data\") pod \"nova-scheduler-0\" (UID: \"6c1fc53e-ad8f-4b72-b1cd-990241b08a43\") " pod="openstack/nova-scheduler-0" Dec 05 11:09:00.628747 master-0 kubenswrapper[24928]: I1205 11:09:00.628683 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c1fc53e-ad8f-4b72-b1cd-990241b08a43-config-data\") pod \"nova-scheduler-0\" (UID: \"6c1fc53e-ad8f-4b72-b1cd-990241b08a43\") " pod="openstack/nova-scheduler-0" Dec 05 11:09:00.628864 master-0 kubenswrapper[24928]: I1205 11:09:00.628736 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c1fc53e-ad8f-4b72-b1cd-990241b08a43-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6c1fc53e-ad8f-4b72-b1cd-990241b08a43\") " pod="openstack/nova-scheduler-0" Dec 05 11:09:00.762503 master-0 kubenswrapper[24928]: I1205 11:09:00.762399 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h85lr\" (UniqueName: \"kubernetes.io/projected/6c1fc53e-ad8f-4b72-b1cd-990241b08a43-kube-api-access-h85lr\") pod \"nova-scheduler-0\" (UID: \"6c1fc53e-ad8f-4b72-b1cd-990241b08a43\") " pod="openstack/nova-scheduler-0" Dec 05 11:09:01.007602 master-0 kubenswrapper[24928]: I1205 11:09:01.007528 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 11:09:01.046817 master-0 kubenswrapper[24928]: I1205 11:09:01.044822 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 11:09:01.053617 master-0 kubenswrapper[24928]: W1205 11:09:01.047390 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod610b724c_d8d9_49e5_9760_cb796b37f0e9.slice/crio-db1250aa78d05b27b838e70be87c52dfc338a6eb65a54d795500b73d5e00f497 WatchSource:0}: Error finding container db1250aa78d05b27b838e70be87c52dfc338a6eb65a54d795500b73d5e00f497: Status 404 returned error can't find the container with id db1250aa78d05b27b838e70be87c52dfc338a6eb65a54d795500b73d5e00f497 Dec 05 11:09:01.228626 master-0 kubenswrapper[24928]: I1205 11:09:01.227787 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62c3336b-a0d7-4e81-b8b2-02bef9bd5984","Type":"ContainerStarted","Data":"e540e7ca1bf9f95b8a368d87edcf03227697d7303f0bda447573ebb610dd6453"} Dec 05 11:09:01.229568 master-0 kubenswrapper[24928]: I1205 11:09:01.228631 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62c3336b-a0d7-4e81-b8b2-02bef9bd5984","Type":"ContainerStarted","Data":"003f3ccf0f9c35abab545c33756fc8cdd08e49b6f1131be7ba51cf7ecd09caf7"} Dec 05 11:09:01.229568 master-0 kubenswrapper[24928]: I1205 11:09:01.228652 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"62c3336b-a0d7-4e81-b8b2-02bef9bd5984","Type":"ContainerStarted","Data":"e7abec0db14632cb2fc2ca9ca479eb10f334960b56000fa6aed3a42728eb1af0"} Dec 05 11:09:01.237397 master-0 kubenswrapper[24928]: I1205 11:09:01.237321 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"610b724c-d8d9-49e5-9760-cb796b37f0e9","Type":"ContainerStarted","Data":"db1250aa78d05b27b838e70be87c52dfc338a6eb65a54d795500b73d5e00f497"} Dec 05 11:09:01.277524 master-0 kubenswrapper[24928]: I1205 11:09:01.277391 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.27736954 podStartE2EDuration="2.27736954s" podCreationTimestamp="2025-12-05 11:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:09:01.264761502 +0000 UTC m=+1301.267955353" watchObservedRunningTime="2025-12-05 11:09:01.27736954 +0000 UTC m=+1301.280563411" Dec 05 11:09:01.524469 master-0 kubenswrapper[24928]: I1205 11:09:01.524319 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 11:09:01.524469 master-0 kubenswrapper[24928]: W1205 11:09:01.524410 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c1fc53e_ad8f_4b72_b1cd_990241b08a43.slice/crio-740dd8be16008623c8e61ecf89d617689b94344e07c198fb95a9cac284aa2835 WatchSource:0}: Error finding container 740dd8be16008623c8e61ecf89d617689b94344e07c198fb95a9cac284aa2835: Status 404 returned error can't find the container with id 740dd8be16008623c8e61ecf89d617689b94344e07c198fb95a9cac284aa2835 Dec 05 11:09:02.219084 master-0 kubenswrapper[24928]: I1205 11:09:02.218994 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42cf2ec9-0902-4907-9b28-8a20dd22a6f0" path="/var/lib/kubelet/pods/42cf2ec9-0902-4907-9b28-8a20dd22a6f0/volumes" Dec 05 11:09:02.258708 master-0 kubenswrapper[24928]: I1205 11:09:02.258604 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"610b724c-d8d9-49e5-9760-cb796b37f0e9","Type":"ContainerStarted","Data":"854100b33b6ca3de67f8c4410de5ebd6bcb7bea4c38f2ba37d23217061222a04"} Dec 05 11:09:02.258708 master-0 kubenswrapper[24928]: I1205 11:09:02.258693 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"610b724c-d8d9-49e5-9760-cb796b37f0e9","Type":"ContainerStarted","Data":"cb458568a9548775cecca7da29120fd7b05afaecc7ddd1ed82d6080584fd195b"} Dec 05 11:09:02.263682 master-0 kubenswrapper[24928]: I1205 11:09:02.263581 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6c1fc53e-ad8f-4b72-b1cd-990241b08a43","Type":"ContainerStarted","Data":"8b813cb38bd4b26b15dadd2a2dbad40ce562c561649ea60a32aa0e3b2acdf6e0"} Dec 05 11:09:02.263786 master-0 kubenswrapper[24928]: I1205 11:09:02.263692 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6c1fc53e-ad8f-4b72-b1cd-990241b08a43","Type":"ContainerStarted","Data":"740dd8be16008623c8e61ecf89d617689b94344e07c198fb95a9cac284aa2835"} Dec 05 11:09:02.307043 master-0 kubenswrapper[24928]: I1205 11:09:02.306797 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.306777288 podStartE2EDuration="3.306777288s" podCreationTimestamp="2025-12-05 11:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:09:02.282942526 +0000 UTC m=+1302.286136397" watchObservedRunningTime="2025-12-05 11:09:02.306777288 +0000 UTC m=+1302.309971139" Dec 05 11:09:05.562215 master-0 kubenswrapper[24928]: I1205 11:09:05.562100 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 11:09:05.562215 master-0 kubenswrapper[24928]: I1205 11:09:05.562216 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 11:09:06.008773 master-0 kubenswrapper[24928]: I1205 11:09:06.008534 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 11:09:09.650373 master-0 kubenswrapper[24928]: I1205 11:09:09.650317 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:09:09.650373 master-0 kubenswrapper[24928]: I1205 11:09:09.650375 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 11:09:10.561662 master-0 kubenswrapper[24928]: I1205 11:09:10.561593 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 11:09:10.561662 master-0 kubenswrapper[24928]: I1205 11:09:10.561662 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 11:09:10.661747 master-0 kubenswrapper[24928]: I1205 11:09:10.661664 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="62c3336b-a0d7-4e81-b8b2-02bef9bd5984" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.128.1.13:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:09:10.662455 master-0 kubenswrapper[24928]: I1205 11:09:10.661664 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="62c3336b-a0d7-4e81-b8b2-02bef9bd5984" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.128.1.13:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:09:11.009654 master-0 kubenswrapper[24928]: I1205 11:09:11.009499 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 11:09:11.056394 master-0 kubenswrapper[24928]: I1205 11:09:11.056321 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 11:09:11.087151 master-0 kubenswrapper[24928]: I1205 11:09:11.087045 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=11.087022422 podStartE2EDuration="11.087022422s" podCreationTimestamp="2025-12-05 11:09:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:09:02.324633933 +0000 UTC m=+1302.327827784" watchObservedRunningTime="2025-12-05 11:09:11.087022422 +0000 UTC m=+1311.090216273" Dec 05 11:09:11.408268 master-0 kubenswrapper[24928]: I1205 11:09:11.407738 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 11:09:11.585453 master-0 kubenswrapper[24928]: I1205 11:09:11.582659 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="610b724c-d8d9-49e5-9760-cb796b37f0e9" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.128.1.14:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:09:11.585453 master-0 kubenswrapper[24928]: I1205 11:09:11.582671 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="610b724c-d8d9-49e5-9760-cb796b37f0e9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.128.1.14:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 11:09:19.658058 master-0 kubenswrapper[24928]: I1205 11:09:19.657748 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 11:09:19.660976 master-0 kubenswrapper[24928]: I1205 11:09:19.660953 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 11:09:19.661658 master-0 kubenswrapper[24928]: I1205 11:09:19.661474 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 11:09:19.668806 master-0 kubenswrapper[24928]: I1205 11:09:19.668763 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 11:09:20.487038 master-0 kubenswrapper[24928]: I1205 11:09:20.486949 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 11:09:20.495505 master-0 kubenswrapper[24928]: I1205 11:09:20.495409 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 11:09:20.577195 master-0 kubenswrapper[24928]: I1205 11:09:20.577134 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 11:09:20.584378 master-0 kubenswrapper[24928]: I1205 11:09:20.578941 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 11:09:20.586731 master-0 kubenswrapper[24928]: I1205 11:09:20.586683 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 11:09:21.503971 master-0 kubenswrapper[24928]: I1205 11:09:21.503911 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 11:09:48.913907 master-0 kubenswrapper[24928]: I1205 11:09:48.913670 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["sushy-emulator/sushy-emulator-58f4c9b998-rkg8l"] Dec 05 11:09:48.914580 master-0 kubenswrapper[24928]: I1205 11:09:48.914081 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" podUID="40a1ca30-6b0d-405a-9626-e23d9ab046b0" containerName="sushy-emulator" containerID="cri-o://ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c" gracePeriod=30 Dec 05 11:09:49.615453 master-0 kubenswrapper[24928]: I1205 11:09:49.615352 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 11:09:49.737604 master-0 kubenswrapper[24928]: I1205 11:09:49.737543 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sushy-emulator-config\" (UniqueName: \"kubernetes.io/configmap/40a1ca30-6b0d-405a-9626-e23d9ab046b0-sushy-emulator-config\") pod \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " Dec 05 11:09:49.737604 master-0 kubenswrapper[24928]: I1205 11:09:49.737611 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"os-client-config\" (UniqueName: \"kubernetes.io/secret/40a1ca30-6b0d-405a-9626-e23d9ab046b0-os-client-config\") pod \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " Dec 05 11:09:49.737962 master-0 kubenswrapper[24928]: I1205 11:09:49.737827 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzzqh\" (UniqueName: \"kubernetes.io/projected/40a1ca30-6b0d-405a-9626-e23d9ab046b0-kube-api-access-dzzqh\") pod \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\" (UID: \"40a1ca30-6b0d-405a-9626-e23d9ab046b0\") " Dec 05 11:09:49.744073 master-0 kubenswrapper[24928]: I1205 11:09:49.744005 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40a1ca30-6b0d-405a-9626-e23d9ab046b0-sushy-emulator-config" (OuterVolumeSpecName: "sushy-emulator-config") pod "40a1ca30-6b0d-405a-9626-e23d9ab046b0" (UID: "40a1ca30-6b0d-405a-9626-e23d9ab046b0"). InnerVolumeSpecName "sushy-emulator-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:09:49.744726 master-0 kubenswrapper[24928]: I1205 11:09:49.744637 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40a1ca30-6b0d-405a-9626-e23d9ab046b0-kube-api-access-dzzqh" (OuterVolumeSpecName: "kube-api-access-dzzqh") pod "40a1ca30-6b0d-405a-9626-e23d9ab046b0" (UID: "40a1ca30-6b0d-405a-9626-e23d9ab046b0"). InnerVolumeSpecName "kube-api-access-dzzqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:09:49.747916 master-0 kubenswrapper[24928]: I1205 11:09:49.747864 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40a1ca30-6b0d-405a-9626-e23d9ab046b0-os-client-config" (OuterVolumeSpecName: "os-client-config") pod "40a1ca30-6b0d-405a-9626-e23d9ab046b0" (UID: "40a1ca30-6b0d-405a-9626-e23d9ab046b0"). InnerVolumeSpecName "os-client-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:09:49.768023 master-0 kubenswrapper[24928]: I1205 11:09:49.766583 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["sushy-emulator/sushy-emulator-64488c485f-p8xkb"] Dec 05 11:09:49.768023 master-0 kubenswrapper[24928]: E1205 11:09:49.767529 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40a1ca30-6b0d-405a-9626-e23d9ab046b0" containerName="sushy-emulator" Dec 05 11:09:49.768023 master-0 kubenswrapper[24928]: I1205 11:09:49.767551 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="40a1ca30-6b0d-405a-9626-e23d9ab046b0" containerName="sushy-emulator" Dec 05 11:09:49.768023 master-0 kubenswrapper[24928]: I1205 11:09:49.767945 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="40a1ca30-6b0d-405a-9626-e23d9ab046b0" containerName="sushy-emulator" Dec 05 11:09:49.768965 master-0 kubenswrapper[24928]: I1205 11:09:49.768926 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["sushy-emulator/sushy-emulator-64488c485f-p8xkb"] Dec 05 11:09:49.769072 master-0 kubenswrapper[24928]: I1205 11:09:49.769043 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:09:49.840869 master-0 kubenswrapper[24928]: I1205 11:09:49.840793 24928 reconciler_common.go:293] "Volume detached for volume \"sushy-emulator-config\" (UniqueName: \"kubernetes.io/configmap/40a1ca30-6b0d-405a-9626-e23d9ab046b0-sushy-emulator-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:09:49.840869 master-0 kubenswrapper[24928]: I1205 11:09:49.840856 24928 reconciler_common.go:293] "Volume detached for volume \"os-client-config\" (UniqueName: \"kubernetes.io/secret/40a1ca30-6b0d-405a-9626-e23d9ab046b0-os-client-config\") on node \"master-0\" DevicePath \"\"" Dec 05 11:09:49.840869 master-0 kubenswrapper[24928]: I1205 11:09:49.840878 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzzqh\" (UniqueName: \"kubernetes.io/projected/40a1ca30-6b0d-405a-9626-e23d9ab046b0-kube-api-access-dzzqh\") on node \"master-0\" DevicePath \"\"" Dec 05 11:09:49.874663 master-0 kubenswrapper[24928]: I1205 11:09:49.874607 24928 generic.go:334] "Generic (PLEG): container finished" podID="40a1ca30-6b0d-405a-9626-e23d9ab046b0" containerID="ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c" exitCode=0 Dec 05 11:09:49.874663 master-0 kubenswrapper[24928]: I1205 11:09:49.874659 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" Dec 05 11:09:49.874959 master-0 kubenswrapper[24928]: I1205 11:09:49.874659 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" event={"ID":"40a1ca30-6b0d-405a-9626-e23d9ab046b0","Type":"ContainerDied","Data":"ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c"} Dec 05 11:09:49.874959 master-0 kubenswrapper[24928]: I1205 11:09:49.874731 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="sushy-emulator/sushy-emulator-58f4c9b998-rkg8l" event={"ID":"40a1ca30-6b0d-405a-9626-e23d9ab046b0","Type":"ContainerDied","Data":"b678a2294c64d0a3cbd335132f0f5a3fb98bb375f75e1455a8b793a7036daf4b"} Dec 05 11:09:49.874959 master-0 kubenswrapper[24928]: I1205 11:09:49.874757 24928 scope.go:117] "RemoveContainer" containerID="ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c" Dec 05 11:09:49.897166 master-0 kubenswrapper[24928]: I1205 11:09:49.897119 24928 scope.go:117] "RemoveContainer" containerID="ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c" Dec 05 11:09:49.897686 master-0 kubenswrapper[24928]: E1205 11:09:49.897645 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c\": container with ID starting with ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c not found: ID does not exist" containerID="ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c" Dec 05 11:09:49.897743 master-0 kubenswrapper[24928]: I1205 11:09:49.897689 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c"} err="failed to get container status \"ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c\": rpc error: code = NotFound desc = could not find container \"ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c\": container with ID starting with ebbd78e12e6737bc57444dc8f2a916adf73a38aa10a9e35f40a6720c28e36e3c not found: ID does not exist" Dec 05 11:09:49.921849 master-0 kubenswrapper[24928]: I1205 11:09:49.921744 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["sushy-emulator/sushy-emulator-58f4c9b998-rkg8l"] Dec 05 11:09:49.936234 master-0 kubenswrapper[24928]: I1205 11:09:49.936149 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["sushy-emulator/sushy-emulator-58f4c9b998-rkg8l"] Dec 05 11:09:49.943124 master-0 kubenswrapper[24928]: I1205 11:09:49.942932 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l28r5\" (UniqueName: \"kubernetes.io/projected/945e21b0-ff3c-482f-a1ea-2b553d9488b1-kube-api-access-l28r5\") pod \"sushy-emulator-64488c485f-p8xkb\" (UID: \"945e21b0-ff3c-482f-a1ea-2b553d9488b1\") " pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:09:49.943481 master-0 kubenswrapper[24928]: I1205 11:09:49.943345 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sushy-emulator-config\" (UniqueName: \"kubernetes.io/configmap/945e21b0-ff3c-482f-a1ea-2b553d9488b1-sushy-emulator-config\") pod \"sushy-emulator-64488c485f-p8xkb\" (UID: \"945e21b0-ff3c-482f-a1ea-2b553d9488b1\") " pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:09:49.943481 master-0 kubenswrapper[24928]: I1205 11:09:49.943435 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-client-config\" (UniqueName: \"kubernetes.io/secret/945e21b0-ff3c-482f-a1ea-2b553d9488b1-os-client-config\") pod \"sushy-emulator-64488c485f-p8xkb\" (UID: \"945e21b0-ff3c-482f-a1ea-2b553d9488b1\") " pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:09:50.046230 master-0 kubenswrapper[24928]: I1205 11:09:50.046137 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-client-config\" (UniqueName: \"kubernetes.io/secret/945e21b0-ff3c-482f-a1ea-2b553d9488b1-os-client-config\") pod \"sushy-emulator-64488c485f-p8xkb\" (UID: \"945e21b0-ff3c-482f-a1ea-2b553d9488b1\") " pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:09:50.046521 master-0 kubenswrapper[24928]: I1205 11:09:50.046468 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l28r5\" (UniqueName: \"kubernetes.io/projected/945e21b0-ff3c-482f-a1ea-2b553d9488b1-kube-api-access-l28r5\") pod \"sushy-emulator-64488c485f-p8xkb\" (UID: \"945e21b0-ff3c-482f-a1ea-2b553d9488b1\") " pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:09:50.046592 master-0 kubenswrapper[24928]: I1205 11:09:50.046532 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sushy-emulator-config\" (UniqueName: \"kubernetes.io/configmap/945e21b0-ff3c-482f-a1ea-2b553d9488b1-sushy-emulator-config\") pod \"sushy-emulator-64488c485f-p8xkb\" (UID: \"945e21b0-ff3c-482f-a1ea-2b553d9488b1\") " pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:09:50.048598 master-0 kubenswrapper[24928]: I1205 11:09:50.048550 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sushy-emulator-config\" (UniqueName: \"kubernetes.io/configmap/945e21b0-ff3c-482f-a1ea-2b553d9488b1-sushy-emulator-config\") pod \"sushy-emulator-64488c485f-p8xkb\" (UID: \"945e21b0-ff3c-482f-a1ea-2b553d9488b1\") " pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:09:50.049856 master-0 kubenswrapper[24928]: I1205 11:09:50.049812 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-client-config\" (UniqueName: \"kubernetes.io/secret/945e21b0-ff3c-482f-a1ea-2b553d9488b1-os-client-config\") pod \"sushy-emulator-64488c485f-p8xkb\" (UID: \"945e21b0-ff3c-482f-a1ea-2b553d9488b1\") " pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:09:50.063881 master-0 kubenswrapper[24928]: I1205 11:09:50.063821 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l28r5\" (UniqueName: \"kubernetes.io/projected/945e21b0-ff3c-482f-a1ea-2b553d9488b1-kube-api-access-l28r5\") pod \"sushy-emulator-64488c485f-p8xkb\" (UID: \"945e21b0-ff3c-482f-a1ea-2b553d9488b1\") " pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:09:50.170516 master-0 kubenswrapper[24928]: I1205 11:09:50.170441 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:09:50.223454 master-0 kubenswrapper[24928]: I1205 11:09:50.221833 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40a1ca30-6b0d-405a-9626-e23d9ab046b0" path="/var/lib/kubelet/pods/40a1ca30-6b0d-405a-9626-e23d9ab046b0/volumes" Dec 05 11:09:50.716228 master-0 kubenswrapper[24928]: W1205 11:09:50.716158 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod945e21b0_ff3c_482f_a1ea_2b553d9488b1.slice/crio-f1f32e1d87ee3458ca5ecfbc0f32da6c3771cebb4c331e9309d6060330a8db9c WatchSource:0}: Error finding container f1f32e1d87ee3458ca5ecfbc0f32da6c3771cebb4c331e9309d6060330a8db9c: Status 404 returned error can't find the container with id f1f32e1d87ee3458ca5ecfbc0f32da6c3771cebb4c331e9309d6060330a8db9c Dec 05 11:09:50.716886 master-0 kubenswrapper[24928]: I1205 11:09:50.716818 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["sushy-emulator/sushy-emulator-64488c485f-p8xkb"] Dec 05 11:09:50.889302 master-0 kubenswrapper[24928]: I1205 11:09:50.889137 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" event={"ID":"945e21b0-ff3c-482f-a1ea-2b553d9488b1","Type":"ContainerStarted","Data":"f1f32e1d87ee3458ca5ecfbc0f32da6c3771cebb4c331e9309d6060330a8db9c"} Dec 05 11:09:51.906499 master-0 kubenswrapper[24928]: I1205 11:09:51.906430 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" event={"ID":"945e21b0-ff3c-482f-a1ea-2b553d9488b1","Type":"ContainerStarted","Data":"70a8484adeda52b9a263294adec7ebeba7dc6a8811b714824761adbd62397475"} Dec 05 11:09:51.953892 master-0 kubenswrapper[24928]: I1205 11:09:51.953807 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" podStartSLOduration=2.953790042 podStartE2EDuration="2.953790042s" podCreationTimestamp="2025-12-05 11:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:09:51.924250351 +0000 UTC m=+1351.927444212" watchObservedRunningTime="2025-12-05 11:09:51.953790042 +0000 UTC m=+1351.956983893" Dec 05 11:10:00.171560 master-0 kubenswrapper[24928]: I1205 11:10:00.171417 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:10:00.171560 master-0 kubenswrapper[24928]: I1205 11:10:00.171549 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:10:00.181439 master-0 kubenswrapper[24928]: I1205 11:10:00.181341 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:10:01.028741 master-0 kubenswrapper[24928]: I1205 11:10:01.028656 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="sushy-emulator/sushy-emulator-64488c485f-p8xkb" Dec 05 11:10:45.461089 master-0 kubenswrapper[24928]: I1205 11:10:45.461021 24928 scope.go:117] "RemoveContainer" containerID="55edfa24d1315fdd7e7862bc5f37f4cce06eb081ab98aa28e20f9d7e06851004" Dec 05 11:10:45.485188 master-0 kubenswrapper[24928]: I1205 11:10:45.485119 24928 scope.go:117] "RemoveContainer" containerID="b1075c17728b1e1aac22815dba6f661231e9ef95c47a4d69f70ff208cad8b81a" Dec 05 11:10:45.552089 master-0 kubenswrapper[24928]: I1205 11:10:45.552023 24928 scope.go:117] "RemoveContainer" containerID="c5fa8195f5b1bf7a6a2263c989136cef1dcda36902f517a5a275bb0a6379a6ee" Dec 05 11:10:45.580746 master-0 kubenswrapper[24928]: I1205 11:10:45.580692 24928 scope.go:117] "RemoveContainer" containerID="37cd0810358922b393a436f10e5de4068a383e25a97be24d99afae8dd338af71" Dec 05 11:10:45.634125 master-0 kubenswrapper[24928]: I1205 11:10:45.634061 24928 scope.go:117] "RemoveContainer" containerID="3a22b1ef4250075eaf588a4cfe46a335f8860c202a5a80b69f54d9361af44736" Dec 05 11:11:45.796340 master-0 kubenswrapper[24928]: I1205 11:11:45.796274 24928 scope.go:117] "RemoveContainer" containerID="8b366a44377a553202afffd5e196805a86d97c265f2d4791b8be0632a6bc7543" Dec 05 11:11:45.829481 master-0 kubenswrapper[24928]: I1205 11:11:45.829414 24928 scope.go:117] "RemoveContainer" containerID="147d4e2b823107dd73c2695e1569f383bc3f459b9c676b6a321be91f112e4b3e" Dec 05 11:12:45.970138 master-0 kubenswrapper[24928]: I1205 11:12:45.970083 24928 scope.go:117] "RemoveContainer" containerID="0bdc22ef7d01cde6d461167b49e502d552500c23ab2d5088432556d87905f364" Dec 05 11:12:45.999514 master-0 kubenswrapper[24928]: I1205 11:12:45.999461 24928 scope.go:117] "RemoveContainer" containerID="48435a82a76dfce0de134e8638b2d8aa1e1168935349387ac394b1432d91dde0" Dec 05 11:12:46.048011 master-0 kubenswrapper[24928]: I1205 11:12:46.047955 24928 scope.go:117] "RemoveContainer" containerID="fa4422211b09287950930e298e844e4ea749f555001e652c0fd0af740e913867" Dec 05 11:13:46.114940 master-0 kubenswrapper[24928]: I1205 11:13:46.114858 24928 scope.go:117] "RemoveContainer" containerID="ebfbe2252b2a4ef8d0f1b9c6c4eb2b0091893ac0af60288bcc8b25ac01c6913c" Dec 05 11:13:46.149522 master-0 kubenswrapper[24928]: I1205 11:13:46.149478 24928 scope.go:117] "RemoveContainer" containerID="6d1b122f166fe65e8e7588f3cd14d0233e19f369beb8056737fd38fc24cd8f3a" Dec 05 11:14:46.249201 master-0 kubenswrapper[24928]: I1205 11:14:46.249122 24928 scope.go:117] "RemoveContainer" containerID="8aab248a4a18b20f40b672910717a747856c3313e10a8afffd34ba7affb7fc60" Dec 05 11:14:46.272664 master-0 kubenswrapper[24928]: I1205 11:14:46.272423 24928 scope.go:117] "RemoveContainer" containerID="9100d422acb02e3be2478df3820b7e706e42eaf5df94ba61f88e701d46253663" Dec 05 11:14:54.136541 master-0 kubenswrapper[24928]: I1205 11:14:54.132526 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-2c49-account-create-update-62tjp"] Dec 05 11:14:54.148167 master-0 kubenswrapper[24928]: I1205 11:14:54.148084 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-2c49-account-create-update-62tjp"] Dec 05 11:14:54.219093 master-0 kubenswrapper[24928]: I1205 11:14:54.219022 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e00a4705-aba4-47f1-8a91-349563989651" path="/var/lib/kubelet/pods/e00a4705-aba4-47f1-8a91-349563989651/volumes" Dec 05 11:14:55.038672 master-0 kubenswrapper[24928]: I1205 11:14:55.038562 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-28pjp"] Dec 05 11:14:55.053996 master-0 kubenswrapper[24928]: I1205 11:14:55.053900 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-28pjp"] Dec 05 11:14:56.221649 master-0 kubenswrapper[24928]: I1205 11:14:56.221492 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3af15ab-5fe0-4536-b61f-77c795f6e7f2" path="/var/lib/kubelet/pods/c3af15ab-5fe0-4536-b61f-77c795f6e7f2/volumes" Dec 05 11:15:00.052246 master-0 kubenswrapper[24928]: I1205 11:15:00.052179 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-tr2vj"] Dec 05 11:15:00.065258 master-0 kubenswrapper[24928]: I1205 11:15:00.065182 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-tr2vj"] Dec 05 11:15:00.174763 master-0 kubenswrapper[24928]: I1205 11:15:00.174364 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8"] Dec 05 11:15:00.178656 master-0 kubenswrapper[24928]: I1205 11:15:00.178564 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:00.182187 master-0 kubenswrapper[24928]: I1205 11:15:00.181953 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-fdxl6" Dec 05 11:15:00.182632 master-0 kubenswrapper[24928]: I1205 11:15:00.182543 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 11:15:00.202989 master-0 kubenswrapper[24928]: I1205 11:15:00.202840 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8"] Dec 05 11:15:00.229588 master-0 kubenswrapper[24928]: I1205 11:15:00.229511 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1ffd92e-85ff-4cab-a0c5-55968a23bc53" path="/var/lib/kubelet/pods/c1ffd92e-85ff-4cab-a0c5-55968a23bc53/volumes" Dec 05 11:15:00.263171 master-0 kubenswrapper[24928]: I1205 11:15:00.263118 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-config-volume\") pod \"collect-profiles-29415555-2kkl8\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:00.263552 master-0 kubenswrapper[24928]: I1205 11:15:00.263240 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-secret-volume\") pod \"collect-profiles-29415555-2kkl8\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:00.263669 master-0 kubenswrapper[24928]: I1205 11:15:00.263625 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c467\" (UniqueName: \"kubernetes.io/projected/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-kube-api-access-7c467\") pod \"collect-profiles-29415555-2kkl8\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:00.366196 master-0 kubenswrapper[24928]: I1205 11:15:00.365979 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-config-volume\") pod \"collect-profiles-29415555-2kkl8\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:00.366196 master-0 kubenswrapper[24928]: I1205 11:15:00.366091 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-secret-volume\") pod \"collect-profiles-29415555-2kkl8\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:00.366535 master-0 kubenswrapper[24928]: I1205 11:15:00.366228 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c467\" (UniqueName: \"kubernetes.io/projected/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-kube-api-access-7c467\") pod \"collect-profiles-29415555-2kkl8\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:00.367403 master-0 kubenswrapper[24928]: I1205 11:15:00.367354 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-config-volume\") pod \"collect-profiles-29415555-2kkl8\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:00.370728 master-0 kubenswrapper[24928]: I1205 11:15:00.370651 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-secret-volume\") pod \"collect-profiles-29415555-2kkl8\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:00.383976 master-0 kubenswrapper[24928]: I1205 11:15:00.383899 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c467\" (UniqueName: \"kubernetes.io/projected/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-kube-api-access-7c467\") pod \"collect-profiles-29415555-2kkl8\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:00.525597 master-0 kubenswrapper[24928]: I1205 11:15:00.525495 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:01.047964 master-0 kubenswrapper[24928]: I1205 11:15:01.047355 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8"] Dec 05 11:15:01.066201 master-0 kubenswrapper[24928]: I1205 11:15:01.066111 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-vb5lw"] Dec 05 11:15:01.081786 master-0 kubenswrapper[24928]: I1205 11:15:01.081696 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-75b5-account-create-update-bpqzn"] Dec 05 11:15:01.103342 master-0 kubenswrapper[24928]: I1205 11:15:01.103270 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-0a14-account-create-update-85wx2"] Dec 05 11:15:01.118592 master-0 kubenswrapper[24928]: I1205 11:15:01.118474 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-vb5lw"] Dec 05 11:15:01.136325 master-0 kubenswrapper[24928]: I1205 11:15:01.136223 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-0a14-account-create-update-85wx2"] Dec 05 11:15:01.152364 master-0 kubenswrapper[24928]: I1205 11:15:01.152224 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-75b5-account-create-update-bpqzn"] Dec 05 11:15:01.785812 master-0 kubenswrapper[24928]: I1205 11:15:01.785695 24928 generic.go:334] "Generic (PLEG): container finished" podID="4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b" containerID="42d0fdf13d0c9c6a6742466868b88381abdb89206977634c217543a730e5abdc" exitCode=0 Dec 05 11:15:01.785812 master-0 kubenswrapper[24928]: I1205 11:15:01.785790 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" event={"ID":"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b","Type":"ContainerDied","Data":"42d0fdf13d0c9c6a6742466868b88381abdb89206977634c217543a730e5abdc"} Dec 05 11:15:01.785812 master-0 kubenswrapper[24928]: I1205 11:15:01.785835 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" event={"ID":"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b","Type":"ContainerStarted","Data":"caf07e5f629d17a3969c33636cf1d77d91a8523f1afa9195ea23dce9eb3696e2"} Dec 05 11:15:02.220217 master-0 kubenswrapper[24928]: I1205 11:15:02.220128 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6592b5de-3923-472c-af99-f7ab0e1caf3f" path="/var/lib/kubelet/pods/6592b5de-3923-472c-af99-f7ab0e1caf3f/volumes" Dec 05 11:15:02.221061 master-0 kubenswrapper[24928]: I1205 11:15:02.220834 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a3249e2-8dd4-42e6-b78e-8712b33d7cf1" path="/var/lib/kubelet/pods/6a3249e2-8dd4-42e6-b78e-8712b33d7cf1/volumes" Dec 05 11:15:02.221477 master-0 kubenswrapper[24928]: I1205 11:15:02.221435 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae071a88-af53-4fb3-93fc-76d8fd311de0" path="/var/lib/kubelet/pods/ae071a88-af53-4fb3-93fc-76d8fd311de0/volumes" Dec 05 11:15:03.253565 master-0 kubenswrapper[24928]: I1205 11:15:03.252657 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:03.348947 master-0 kubenswrapper[24928]: I1205 11:15:03.348880 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-secret-volume\") pod \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " Dec 05 11:15:03.348947 master-0 kubenswrapper[24928]: I1205 11:15:03.348934 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-config-volume\") pod \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " Dec 05 11:15:03.349658 master-0 kubenswrapper[24928]: I1205 11:15:03.349037 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c467\" (UniqueName: \"kubernetes.io/projected/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-kube-api-access-7c467\") pod \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\" (UID: \"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b\") " Dec 05 11:15:03.349658 master-0 kubenswrapper[24928]: I1205 11:15:03.349549 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-config-volume" (OuterVolumeSpecName: "config-volume") pod "4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b" (UID: "4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:15:03.349962 master-0 kubenswrapper[24928]: I1205 11:15:03.349926 24928 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-config-volume\") on node \"master-0\" DevicePath \"\"" Dec 05 11:15:03.351845 master-0 kubenswrapper[24928]: I1205 11:15:03.351802 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b" (UID: "4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:15:03.352154 master-0 kubenswrapper[24928]: I1205 11:15:03.352012 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-kube-api-access-7c467" (OuterVolumeSpecName: "kube-api-access-7c467") pod "4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b" (UID: "4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b"). InnerVolumeSpecName "kube-api-access-7c467". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:15:03.451858 master-0 kubenswrapper[24928]: I1205 11:15:03.451784 24928 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-secret-volume\") on node \"master-0\" DevicePath \"\"" Dec 05 11:15:03.451858 master-0 kubenswrapper[24928]: I1205 11:15:03.451836 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c467\" (UniqueName: \"kubernetes.io/projected/4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b-kube-api-access-7c467\") on node \"master-0\" DevicePath \"\"" Dec 05 11:15:03.809303 master-0 kubenswrapper[24928]: I1205 11:15:03.809241 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" event={"ID":"4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b","Type":"ContainerDied","Data":"caf07e5f629d17a3969c33636cf1d77d91a8523f1afa9195ea23dce9eb3696e2"} Dec 05 11:15:03.809303 master-0 kubenswrapper[24928]: I1205 11:15:03.809300 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caf07e5f629d17a3969c33636cf1d77d91a8523f1afa9195ea23dce9eb3696e2" Dec 05 11:15:03.809594 master-0 kubenswrapper[24928]: I1205 11:15:03.809348 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415555-2kkl8" Dec 05 11:15:28.075633 master-0 kubenswrapper[24928]: I1205 11:15:28.075520 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-5m2fn"] Dec 05 11:15:28.089214 master-0 kubenswrapper[24928]: I1205 11:15:28.089122 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-5m2fn"] Dec 05 11:15:28.217961 master-0 kubenswrapper[24928]: I1205 11:15:28.217889 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1af478de-35a0-4480-a9cd-a2fa1a7ab774" path="/var/lib/kubelet/pods/1af478de-35a0-4480-a9cd-a2fa1a7ab774/volumes" Dec 05 11:15:29.056353 master-0 kubenswrapper[24928]: I1205 11:15:29.056269 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-kf7cj"] Dec 05 11:15:29.068328 master-0 kubenswrapper[24928]: I1205 11:15:29.068224 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d29-account-create-update-dz2qv"] Dec 05 11:15:29.081684 master-0 kubenswrapper[24928]: I1205 11:15:29.081598 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-72f7-account-create-update-rb6rq"] Dec 05 11:15:29.094357 master-0 kubenswrapper[24928]: I1205 11:15:29.094276 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5d29-account-create-update-dz2qv"] Dec 05 11:15:29.105180 master-0 kubenswrapper[24928]: I1205 11:15:29.105117 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-kf7cj"] Dec 05 11:15:29.116503 master-0 kubenswrapper[24928]: I1205 11:15:29.116396 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-72f7-account-create-update-rb6rq"] Dec 05 11:15:30.226936 master-0 kubenswrapper[24928]: I1205 11:15:30.226713 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0da83455-cb84-45a3-bbe7-211be81f4b56" path="/var/lib/kubelet/pods/0da83455-cb84-45a3-bbe7-211be81f4b56/volumes" Dec 05 11:15:30.227913 master-0 kubenswrapper[24928]: I1205 11:15:30.227873 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17f4e63e-29f0-4884-a20e-a3cb02c85235" path="/var/lib/kubelet/pods/17f4e63e-29f0-4884-a20e-a3cb02c85235/volumes" Dec 05 11:15:30.229136 master-0 kubenswrapper[24928]: I1205 11:15:30.229034 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c3b42bc-736b-4910-9bef-39ebeba0c5dd" path="/var/lib/kubelet/pods/1c3b42bc-736b-4910-9bef-39ebeba0c5dd/volumes" Dec 05 11:15:31.058650 master-0 kubenswrapper[24928]: I1205 11:15:31.058583 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fttsk"] Dec 05 11:15:31.059210 master-0 kubenswrapper[24928]: E1205 11:15:31.059182 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b" containerName="collect-profiles" Dec 05 11:15:31.059210 master-0 kubenswrapper[24928]: I1205 11:15:31.059208 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b" containerName="collect-profiles" Dec 05 11:15:31.059601 master-0 kubenswrapper[24928]: I1205 11:15:31.059578 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d77b3f3-78f1-46ff-8d1e-51f4aa2fbd0b" containerName="collect-profiles" Dec 05 11:15:31.061887 master-0 kubenswrapper[24928]: I1205 11:15:31.061843 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:31.094545 master-0 kubenswrapper[24928]: I1205 11:15:31.089900 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fttsk"] Dec 05 11:15:31.124732 master-0 kubenswrapper[24928]: I1205 11:15:31.124652 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x4kb\" (UniqueName: \"kubernetes.io/projected/b15005f0-2323-46d7-916c-473b2443a139-kube-api-access-2x4kb\") pod \"certified-operators-fttsk\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:31.124732 master-0 kubenswrapper[24928]: I1205 11:15:31.124726 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-catalog-content\") pod \"certified-operators-fttsk\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:31.125025 master-0 kubenswrapper[24928]: I1205 11:15:31.124762 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-utilities\") pod \"certified-operators-fttsk\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:31.229822 master-0 kubenswrapper[24928]: I1205 11:15:31.228711 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x4kb\" (UniqueName: \"kubernetes.io/projected/b15005f0-2323-46d7-916c-473b2443a139-kube-api-access-2x4kb\") pod \"certified-operators-fttsk\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:31.229822 master-0 kubenswrapper[24928]: I1205 11:15:31.228778 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-catalog-content\") pod \"certified-operators-fttsk\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:31.229822 master-0 kubenswrapper[24928]: I1205 11:15:31.229453 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-utilities\") pod \"certified-operators-fttsk\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:31.243215 master-0 kubenswrapper[24928]: I1205 11:15:31.229966 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-utilities\") pod \"certified-operators-fttsk\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:31.243215 master-0 kubenswrapper[24928]: I1205 11:15:31.229979 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-catalog-content\") pod \"certified-operators-fttsk\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:31.251265 master-0 kubenswrapper[24928]: I1205 11:15:31.251194 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x4kb\" (UniqueName: \"kubernetes.io/projected/b15005f0-2323-46d7-916c-473b2443a139-kube-api-access-2x4kb\") pod \"certified-operators-fttsk\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:31.394485 master-0 kubenswrapper[24928]: I1205 11:15:31.394314 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:31.965829 master-0 kubenswrapper[24928]: I1205 11:15:31.965763 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fttsk"] Dec 05 11:15:32.134836 master-0 kubenswrapper[24928]: I1205 11:15:32.134756 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fttsk" event={"ID":"b15005f0-2323-46d7-916c-473b2443a139","Type":"ContainerStarted","Data":"fb80f98b5e24f7177fa7ab53069a744824a0409a95dcfd70956edebe0bdc5f57"} Dec 05 11:15:33.156293 master-0 kubenswrapper[24928]: I1205 11:15:33.156229 24928 generic.go:334] "Generic (PLEG): container finished" podID="b15005f0-2323-46d7-916c-473b2443a139" containerID="000219dbe5e2f2ec33676c0f9f396953734f000d042e38162f04c3813c67e66d" exitCode=0 Dec 05 11:15:33.156293 master-0 kubenswrapper[24928]: I1205 11:15:33.156294 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fttsk" event={"ID":"b15005f0-2323-46d7-916c-473b2443a139","Type":"ContainerDied","Data":"000219dbe5e2f2ec33676c0f9f396953734f000d042e38162f04c3813c67e66d"} Dec 05 11:15:33.158738 master-0 kubenswrapper[24928]: I1205 11:15:33.158588 24928 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:15:34.170065 master-0 kubenswrapper[24928]: I1205 11:15:34.169919 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fttsk" event={"ID":"b15005f0-2323-46d7-916c-473b2443a139","Type":"ContainerStarted","Data":"a7c3812fc896405c212b535e766dcb4bec62bbcae1dab944dd16f9441a64d92f"} Dec 05 11:15:35.181910 master-0 kubenswrapper[24928]: I1205 11:15:35.181832 24928 generic.go:334] "Generic (PLEG): container finished" podID="b15005f0-2323-46d7-916c-473b2443a139" containerID="a7c3812fc896405c212b535e766dcb4bec62bbcae1dab944dd16f9441a64d92f" exitCode=0 Dec 05 11:15:35.181910 master-0 kubenswrapper[24928]: I1205 11:15:35.181895 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fttsk" event={"ID":"b15005f0-2323-46d7-916c-473b2443a139","Type":"ContainerDied","Data":"a7c3812fc896405c212b535e766dcb4bec62bbcae1dab944dd16f9441a64d92f"} Dec 05 11:15:36.051915 master-0 kubenswrapper[24928]: I1205 11:15:36.051844 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-bp44h"] Dec 05 11:15:36.065438 master-0 kubenswrapper[24928]: I1205 11:15:36.065353 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-bp44h"] Dec 05 11:15:36.194706 master-0 kubenswrapper[24928]: I1205 11:15:36.194636 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fttsk" event={"ID":"b15005f0-2323-46d7-916c-473b2443a139","Type":"ContainerStarted","Data":"239ad76b0030d21e74d8ab7a8f4aaa25efbf0e03b2f3e5afee3271df9287a4c4"} Dec 05 11:15:36.219205 master-0 kubenswrapper[24928]: I1205 11:15:36.219139 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d10625d-a13a-4538-9d41-6e71f1af3669" path="/var/lib/kubelet/pods/1d10625d-a13a-4538-9d41-6e71f1af3669/volumes" Dec 05 11:15:36.228462 master-0 kubenswrapper[24928]: I1205 11:15:36.227943 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fttsk" podStartSLOduration=2.682306273 podStartE2EDuration="5.227923292s" podCreationTimestamp="2025-12-05 11:15:31 +0000 UTC" firstStartedPulling="2025-12-05 11:15:33.158490962 +0000 UTC m=+1693.161684813" lastFinishedPulling="2025-12-05 11:15:35.704107981 +0000 UTC m=+1695.707301832" observedRunningTime="2025-12-05 11:15:36.216247317 +0000 UTC m=+1696.219441168" watchObservedRunningTime="2025-12-05 11:15:36.227923292 +0000 UTC m=+1696.231117143" Dec 05 11:15:39.041190 master-0 kubenswrapper[24928]: I1205 11:15:39.041117 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-n74vh"] Dec 05 11:15:39.053455 master-0 kubenswrapper[24928]: I1205 11:15:39.053358 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-n74vh"] Dec 05 11:15:40.219952 master-0 kubenswrapper[24928]: I1205 11:15:40.219860 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f038c921-52be-46ef-ad30-190252e42dcb" path="/var/lib/kubelet/pods/f038c921-52be-46ef-ad30-190252e42dcb/volumes" Dec 05 11:15:41.395363 master-0 kubenswrapper[24928]: I1205 11:15:41.395291 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:41.395909 master-0 kubenswrapper[24928]: I1205 11:15:41.395377 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:41.453349 master-0 kubenswrapper[24928]: I1205 11:15:41.453236 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:42.337046 master-0 kubenswrapper[24928]: I1205 11:15:42.336926 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:42.425477 master-0 kubenswrapper[24928]: I1205 11:15:42.423998 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fttsk"] Dec 05 11:15:44.300059 master-0 kubenswrapper[24928]: I1205 11:15:44.299951 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fttsk" podUID="b15005f0-2323-46d7-916c-473b2443a139" containerName="registry-server" containerID="cri-o://239ad76b0030d21e74d8ab7a8f4aaa25efbf0e03b2f3e5afee3271df9287a4c4" gracePeriod=2 Dec 05 11:15:45.312883 master-0 kubenswrapper[24928]: I1205 11:15:45.312827 24928 generic.go:334] "Generic (PLEG): container finished" podID="b15005f0-2323-46d7-916c-473b2443a139" containerID="239ad76b0030d21e74d8ab7a8f4aaa25efbf0e03b2f3e5afee3271df9287a4c4" exitCode=0 Dec 05 11:15:45.312883 master-0 kubenswrapper[24928]: I1205 11:15:45.312880 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fttsk" event={"ID":"b15005f0-2323-46d7-916c-473b2443a139","Type":"ContainerDied","Data":"239ad76b0030d21e74d8ab7a8f4aaa25efbf0e03b2f3e5afee3271df9287a4c4"} Dec 05 11:15:45.867683 master-0 kubenswrapper[24928]: I1205 11:15:45.867626 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:45.961297 master-0 kubenswrapper[24928]: I1205 11:15:45.961224 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x4kb\" (UniqueName: \"kubernetes.io/projected/b15005f0-2323-46d7-916c-473b2443a139-kube-api-access-2x4kb\") pod \"b15005f0-2323-46d7-916c-473b2443a139\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " Dec 05 11:15:45.961297 master-0 kubenswrapper[24928]: I1205 11:15:45.961303 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-utilities\") pod \"b15005f0-2323-46d7-916c-473b2443a139\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " Dec 05 11:15:45.961598 master-0 kubenswrapper[24928]: I1205 11:15:45.961429 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-catalog-content\") pod \"b15005f0-2323-46d7-916c-473b2443a139\" (UID: \"b15005f0-2323-46d7-916c-473b2443a139\") " Dec 05 11:15:45.962642 master-0 kubenswrapper[24928]: I1205 11:15:45.962571 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-utilities" (OuterVolumeSpecName: "utilities") pod "b15005f0-2323-46d7-916c-473b2443a139" (UID: "b15005f0-2323-46d7-916c-473b2443a139"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:15:45.964579 master-0 kubenswrapper[24928]: I1205 11:15:45.964525 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b15005f0-2323-46d7-916c-473b2443a139-kube-api-access-2x4kb" (OuterVolumeSpecName: "kube-api-access-2x4kb") pod "b15005f0-2323-46d7-916c-473b2443a139" (UID: "b15005f0-2323-46d7-916c-473b2443a139"). InnerVolumeSpecName "kube-api-access-2x4kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:15:46.015221 master-0 kubenswrapper[24928]: I1205 11:15:46.015072 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b15005f0-2323-46d7-916c-473b2443a139" (UID: "b15005f0-2323-46d7-916c-473b2443a139"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:15:46.065284 master-0 kubenswrapper[24928]: I1205 11:15:46.064826 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:15:46.065284 master-0 kubenswrapper[24928]: I1205 11:15:46.064878 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x4kb\" (UniqueName: \"kubernetes.io/projected/b15005f0-2323-46d7-916c-473b2443a139-kube-api-access-2x4kb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:15:46.065284 master-0 kubenswrapper[24928]: I1205 11:15:46.064892 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b15005f0-2323-46d7-916c-473b2443a139-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:15:46.117406 master-0 kubenswrapper[24928]: I1205 11:15:46.117253 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-db-create-x99hv"] Dec 05 11:15:46.134213 master-0 kubenswrapper[24928]: I1205 11:15:46.134111 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-db-create-x99hv"] Dec 05 11:15:46.221577 master-0 kubenswrapper[24928]: I1205 11:15:46.221514 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="415910fe-1441-44d5-a220-322e3ba3e2e3" path="/var/lib/kubelet/pods/415910fe-1441-44d5-a220-322e3ba3e2e3/volumes" Dec 05 11:15:46.328853 master-0 kubenswrapper[24928]: I1205 11:15:46.328800 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fttsk" event={"ID":"b15005f0-2323-46d7-916c-473b2443a139","Type":"ContainerDied","Data":"fb80f98b5e24f7177fa7ab53069a744824a0409a95dcfd70956edebe0bdc5f57"} Dec 05 11:15:46.329515 master-0 kubenswrapper[24928]: I1205 11:15:46.329497 24928 scope.go:117] "RemoveContainer" containerID="239ad76b0030d21e74d8ab7a8f4aaa25efbf0e03b2f3e5afee3271df9287a4c4" Dec 05 11:15:46.329714 master-0 kubenswrapper[24928]: I1205 11:15:46.328902 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fttsk" Dec 05 11:15:46.333840 master-0 kubenswrapper[24928]: I1205 11:15:46.333785 24928 scope.go:117] "RemoveContainer" containerID="fc404e48fb0c4ef209a261dfd42695a5705750c1cc8ff2b11f16be3e127c429e" Dec 05 11:15:46.365562 master-0 kubenswrapper[24928]: I1205 11:15:46.365519 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fttsk"] Dec 05 11:15:46.377554 master-0 kubenswrapper[24928]: I1205 11:15:46.377485 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fttsk"] Dec 05 11:15:46.384831 master-0 kubenswrapper[24928]: I1205 11:15:46.384572 24928 scope.go:117] "RemoveContainer" containerID="a7c3812fc896405c212b535e766dcb4bec62bbcae1dab944dd16f9441a64d92f" Dec 05 11:15:46.402873 master-0 kubenswrapper[24928]: I1205 11:15:46.402763 24928 scope.go:117] "RemoveContainer" containerID="d97649f4a534cd578387ebeee2390d3bff81593747f01efc9601dd13d8217795" Dec 05 11:15:46.479512 master-0 kubenswrapper[24928]: I1205 11:15:46.479463 24928 scope.go:117] "RemoveContainer" containerID="000219dbe5e2f2ec33676c0f9f396953734f000d042e38162f04c3813c67e66d" Dec 05 11:15:46.493562 master-0 kubenswrapper[24928]: I1205 11:15:46.493334 24928 scope.go:117] "RemoveContainer" containerID="0c8d87e172238ddbcec7bd55d985c51a25f017fae7c7d331c80fbbbad8c2b0f7" Dec 05 11:15:46.577101 master-0 kubenswrapper[24928]: I1205 11:15:46.577039 24928 scope.go:117] "RemoveContainer" containerID="8b6b32e3398efe0eb9fdc0d0519779f30f19812a641a6dead4c0ffda77141b77" Dec 05 11:15:46.615138 master-0 kubenswrapper[24928]: I1205 11:15:46.614979 24928 scope.go:117] "RemoveContainer" containerID="1cebad37928253659fa3e85f9d986ce4e48e869f5a6564967bb363fec503cb1d" Dec 05 11:15:46.641875 master-0 kubenswrapper[24928]: I1205 11:15:46.641824 24928 scope.go:117] "RemoveContainer" containerID="b38c6de7050c6c6b24d2ae10f75e1b7c84de9556230059b512e58349b6042964" Dec 05 11:15:46.669563 master-0 kubenswrapper[24928]: I1205 11:15:46.669511 24928 scope.go:117] "RemoveContainer" containerID="bdcd9e7bb6583c8e0cbcf590d5ffdb517083382508c79285da8340efa4efe568" Dec 05 11:15:46.692791 master-0 kubenswrapper[24928]: I1205 11:15:46.692749 24928 scope.go:117] "RemoveContainer" containerID="01008e247454c1f7b3a1ca02c81f438c58d10f34ab803c5c6a2b14645ce5b3c2" Dec 05 11:15:46.714461 master-0 kubenswrapper[24928]: I1205 11:15:46.714385 24928 scope.go:117] "RemoveContainer" containerID="6c94531a9e45f78a5382d174371e1bcb6e1cbbfbbca55c1ca06780e0f1c6d491" Dec 05 11:15:46.738127 master-0 kubenswrapper[24928]: I1205 11:15:46.737745 24928 scope.go:117] "RemoveContainer" containerID="3370f573d17f498b4ea25781e741647ab3a6193bb5b0f8a6e64abaa22c3dab7d" Dec 05 11:15:46.787461 master-0 kubenswrapper[24928]: I1205 11:15:46.787368 24928 scope.go:117] "RemoveContainer" containerID="e2f6e1eb2ed9d64ace14666de64091f35b9f6f07353f42395136dc8d33792453" Dec 05 11:15:46.811726 master-0 kubenswrapper[24928]: I1205 11:15:46.811667 24928 scope.go:117] "RemoveContainer" containerID="18d2f406d5d95ce2b4693cd7c1117c8b7d5bd0491c94e207cd7060fae9bbd22d" Dec 05 11:15:46.837229 master-0 kubenswrapper[24928]: I1205 11:15:46.837154 24928 scope.go:117] "RemoveContainer" containerID="af05b34019d940a91a41a72ec26ef13b704ff3e06c119a443966f570240d62c2" Dec 05 11:15:48.105285 master-0 kubenswrapper[24928]: E1205 11:15:48.105234 24928 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 192.168.32.10:59960->192.168.32.10:40581: write tcp 192.168.32.10:59960->192.168.32.10:40581: write: broken pipe Dec 05 11:15:48.220746 master-0 kubenswrapper[24928]: I1205 11:15:48.220659 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b15005f0-2323-46d7-916c-473b2443a139" path="/var/lib/kubelet/pods/b15005f0-2323-46d7-916c-473b2443a139/volumes" Dec 05 11:15:49.118061 master-0 kubenswrapper[24928]: I1205 11:15:49.117944 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-ac30-account-create-update-ngf65"] Dec 05 11:15:49.138107 master-0 kubenswrapper[24928]: I1205 11:15:49.137999 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-ac30-account-create-update-ngf65"] Dec 05 11:15:50.216803 master-0 kubenswrapper[24928]: I1205 11:15:50.216735 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84942dea-64c3-403d-9d76-bbc2d9206f9f" path="/var/lib/kubelet/pods/84942dea-64c3-403d-9d76-bbc2d9206f9f/volumes" Dec 05 11:16:06.063399 master-0 kubenswrapper[24928]: I1205 11:16:06.063321 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-rrvhc"] Dec 05 11:16:06.076923 master-0 kubenswrapper[24928]: I1205 11:16:06.076826 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-rrvhc"] Dec 05 11:16:06.222018 master-0 kubenswrapper[24928]: I1205 11:16:06.221961 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd047ad1-2e39-47f6-ad8f-97eb080e7766" path="/var/lib/kubelet/pods/cd047ad1-2e39-47f6-ad8f-97eb080e7766/volumes" Dec 05 11:16:10.035559 master-0 kubenswrapper[24928]: I1205 11:16:10.035492 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2wx5d"] Dec 05 11:16:10.048567 master-0 kubenswrapper[24928]: I1205 11:16:10.048492 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2wx5d"] Dec 05 11:16:10.220703 master-0 kubenswrapper[24928]: I1205 11:16:10.220636 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11a17674-7ccb-402d-9910-7ec8797f38b8" path="/var/lib/kubelet/pods/11a17674-7ccb-402d-9910-7ec8797f38b8/volumes" Dec 05 11:16:15.187768 master-0 kubenswrapper[24928]: I1205 11:16:15.187692 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ec941-db-sync-9nbmc"] Dec 05 11:16:15.206276 master-0 kubenswrapper[24928]: I1205 11:16:15.204594 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-ec941-db-sync-9nbmc"] Dec 05 11:16:16.219404 master-0 kubenswrapper[24928]: I1205 11:16:16.219315 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ae72689-6505-4064-bd26-861bda2f68cc" path="/var/lib/kubelet/pods/4ae72689-6505-4064-bd26-861bda2f68cc/volumes" Dec 05 11:16:17.056436 master-0 kubenswrapper[24928]: I1205 11:16:17.056327 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-djbqg"] Dec 05 11:16:17.070046 master-0 kubenswrapper[24928]: I1205 11:16:17.069923 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-djbqg"] Dec 05 11:16:18.220123 master-0 kubenswrapper[24928]: I1205 11:16:18.220045 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90315296-e6a2-4db3-aa3d-35af0d21a55e" path="/var/lib/kubelet/pods/90315296-e6a2-4db3-aa3d-35af0d21a55e/volumes" Dec 05 11:16:34.056016 master-0 kubenswrapper[24928]: I1205 11:16:34.055916 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-db-sync-sxns9"] Dec 05 11:16:34.072249 master-0 kubenswrapper[24928]: I1205 11:16:34.072184 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-db-sync-sxns9"] Dec 05 11:16:34.235149 master-0 kubenswrapper[24928]: I1205 11:16:34.235081 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e18fdd2-be67-4228-82a3-4d02d3b350cf" path="/var/lib/kubelet/pods/0e18fdd2-be67-4228-82a3-4d02d3b350cf/volumes" Dec 05 11:16:43.121827 master-0 kubenswrapper[24928]: I1205 11:16:43.121732 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-b0db-account-create-update-5x72f"] Dec 05 11:16:43.140143 master-0 kubenswrapper[24928]: I1205 11:16:43.140061 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-db-create-z4xfk"] Dec 05 11:16:43.159127 master-0 kubenswrapper[24928]: I1205 11:16:43.159014 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-db-create-z4xfk"] Dec 05 11:16:43.177210 master-0 kubenswrapper[24928]: I1205 11:16:43.177130 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-b0db-account-create-update-5x72f"] Dec 05 11:16:44.219328 master-0 kubenswrapper[24928]: I1205 11:16:44.219271 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa" path="/var/lib/kubelet/pods/55a45c7f-83ce-4116-bb51-8e5cb4dfb4fa/volumes" Dec 05 11:16:44.220019 master-0 kubenswrapper[24928]: I1205 11:16:44.219994 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb7241f8-33c3-4dde-95c8-41a315d34ec8" path="/var/lib/kubelet/pods/eb7241f8-33c3-4dde-95c8-41a315d34ec8/volumes" Dec 05 11:16:47.164224 master-0 kubenswrapper[24928]: I1205 11:16:47.164134 24928 scope.go:117] "RemoveContainer" containerID="349e2e75b98f5ec0c4eb43ce361e97dfbc1848d82e597f7be07ffc8be044fa44" Dec 05 11:16:47.212587 master-0 kubenswrapper[24928]: I1205 11:16:47.212526 24928 scope.go:117] "RemoveContainer" containerID="d9d3a6f3db4e457acdf2eb32316c388679088a54bd3b8dee9f35a85b238f51c1" Dec 05 11:16:47.266157 master-0 kubenswrapper[24928]: I1205 11:16:47.265785 24928 scope.go:117] "RemoveContainer" containerID="d3b829955ca0a2f950c80b244ce91bc30a45fc6ecffaf9ceea3e6d188ed8be3c" Dec 05 11:16:47.325869 master-0 kubenswrapper[24928]: I1205 11:16:47.325733 24928 scope.go:117] "RemoveContainer" containerID="cd6f3e3faf8ea4c6c8e88243dcd954d362933320b1173e42a2be96d848c1dff2" Dec 05 11:16:47.382637 master-0 kubenswrapper[24928]: I1205 11:16:47.382567 24928 scope.go:117] "RemoveContainer" containerID="a33daad2110904d887bdef13f8dcdbc0e41b61ab952bb4b352f57727fbcd6e6e" Dec 05 11:16:47.453240 master-0 kubenswrapper[24928]: I1205 11:16:47.453196 24928 scope.go:117] "RemoveContainer" containerID="0fd5eac4fad287c65188ecc90c8530f87659bfbe84a9ec67de0571e498c18967" Dec 05 11:16:47.503351 master-0 kubenswrapper[24928]: I1205 11:16:47.502068 24928 scope.go:117] "RemoveContainer" containerID="f599b091f4901fcc1add1364c893008e0c0791f7fd67d295d56483df969df21d" Dec 05 11:16:47.533652 master-0 kubenswrapper[24928]: I1205 11:16:47.533582 24928 scope.go:117] "RemoveContainer" containerID="612a89b78524e8e7ec8f1b6f48f4864f9f31a47c6deaf2794d8e41c671c1fc6f" Dec 05 11:16:47.569769 master-0 kubenswrapper[24928]: I1205 11:16:47.569716 24928 scope.go:117] "RemoveContainer" containerID="faa12fcdcf2ef663a50455c97db88d084a0ba62848d5c28f9330b220ff68e32d" Dec 05 11:17:12.071929 master-0 kubenswrapper[24928]: I1205 11:17:12.071842 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-db-sync-sppfx"] Dec 05 11:17:12.084331 master-0 kubenswrapper[24928]: I1205 11:17:12.084111 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-db-sync-sppfx"] Dec 05 11:17:12.233545 master-0 kubenswrapper[24928]: I1205 11:17:12.233409 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5183387d-f79b-4412-88ac-41b37866e07e" path="/var/lib/kubelet/pods/5183387d-f79b-4412-88ac-41b37866e07e/volumes" Dec 05 11:17:23.046003 master-0 kubenswrapper[24928]: I1205 11:17:23.045937 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-tqgcv"] Dec 05 11:17:23.056738 master-0 kubenswrapper[24928]: I1205 11:17:23.056666 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-1c49-account-create-update-vwpvg"] Dec 05 11:17:23.067099 master-0 kubenswrapper[24928]: I1205 11:17:23.067023 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-tqgcv"] Dec 05 11:17:23.078362 master-0 kubenswrapper[24928]: I1205 11:17:23.078293 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-9b34-account-create-update-gllvm"] Dec 05 11:17:23.089009 master-0 kubenswrapper[24928]: I1205 11:17:23.088949 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-g2gnb"] Dec 05 11:17:23.101935 master-0 kubenswrapper[24928]: I1205 11:17:23.101865 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-1c49-account-create-update-vwpvg"] Dec 05 11:17:23.113187 master-0 kubenswrapper[24928]: I1205 11:17:23.113095 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-649rk"] Dec 05 11:17:23.127948 master-0 kubenswrapper[24928]: I1205 11:17:23.127865 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-9b34-account-create-update-gllvm"] Dec 05 11:17:23.139215 master-0 kubenswrapper[24928]: I1205 11:17:23.139116 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-g2gnb"] Dec 05 11:17:23.150167 master-0 kubenswrapper[24928]: I1205 11:17:23.149994 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-649rk"] Dec 05 11:17:24.047177 master-0 kubenswrapper[24928]: I1205 11:17:24.047096 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-2b0e-account-create-update-fpn9k"] Dec 05 11:17:24.058002 master-0 kubenswrapper[24928]: I1205 11:17:24.057884 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-2b0e-account-create-update-fpn9k"] Dec 05 11:17:24.223535 master-0 kubenswrapper[24928]: I1205 11:17:24.223309 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06c47656-d5b2-405d-944d-3d4ce03be7b0" path="/var/lib/kubelet/pods/06c47656-d5b2-405d-944d-3d4ce03be7b0/volumes" Dec 05 11:17:24.224894 master-0 kubenswrapper[24928]: I1205 11:17:24.224854 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82b65c14-8968-437d-840d-30ccd28c57f9" path="/var/lib/kubelet/pods/82b65c14-8968-437d-840d-30ccd28c57f9/volumes" Dec 05 11:17:24.225610 master-0 kubenswrapper[24928]: I1205 11:17:24.225579 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83a55bec-0267-4de3-97e8-8705a6056565" path="/var/lib/kubelet/pods/83a55bec-0267-4de3-97e8-8705a6056565/volumes" Dec 05 11:17:24.226242 master-0 kubenswrapper[24928]: I1205 11:17:24.226212 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c86a407-7dbd-49d1-87e8-c447bceb67bb" path="/var/lib/kubelet/pods/9c86a407-7dbd-49d1-87e8-c447bceb67bb/volumes" Dec 05 11:17:24.227495 master-0 kubenswrapper[24928]: I1205 11:17:24.227383 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e73ed846-11a8-4c4b-b779-d7d362a3933f" path="/var/lib/kubelet/pods/e73ed846-11a8-4c4b-b779-d7d362a3933f/volumes" Dec 05 11:17:24.228033 master-0 kubenswrapper[24928]: I1205 11:17:24.227977 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f180288f-009b-4f64-bf07-543726ebc2e2" path="/var/lib/kubelet/pods/f180288f-009b-4f64-bf07-543726ebc2e2/volumes" Dec 05 11:17:47.742782 master-0 kubenswrapper[24928]: I1205 11:17:47.742711 24928 scope.go:117] "RemoveContainer" containerID="aa96d63f095df2ca19ce1477b28f29c86bdcdf74b8f2261ad3dcd4bc3f93b052" Dec 05 11:17:47.770811 master-0 kubenswrapper[24928]: I1205 11:17:47.769801 24928 scope.go:117] "RemoveContainer" containerID="9fb6867aa7cfa7925e04c045e72c2da1fe1c373a0b7241f16eb9a30adbcfcf11" Dec 05 11:17:47.863322 master-0 kubenswrapper[24928]: I1205 11:17:47.863267 24928 scope.go:117] "RemoveContainer" containerID="4e24ad44d524026003f4888853d5d3caeac4e3f2ac890101297980e102d96974" Dec 05 11:17:47.903440 master-0 kubenswrapper[24928]: I1205 11:17:47.903374 24928 scope.go:117] "RemoveContainer" containerID="0b891267e1e08f06ff8f2233654172342f8078c9cf3a0e2fc711005a5b723ba8" Dec 05 11:17:47.954105 master-0 kubenswrapper[24928]: I1205 11:17:47.954034 24928 scope.go:117] "RemoveContainer" containerID="fd6274db687f6976b1c97618655aae11808c055a7cad0201a5eb5b0af9605762" Dec 05 11:17:48.011681 master-0 kubenswrapper[24928]: I1205 11:17:48.011624 24928 scope.go:117] "RemoveContainer" containerID="acce575d949e3045ed4bd69b5f14b9260c305ebba6ba3756fcaf6c5db9f0ad61" Dec 05 11:17:48.055317 master-0 kubenswrapper[24928]: I1205 11:17:48.055258 24928 scope.go:117] "RemoveContainer" containerID="f459c0129b9cd0a94447984332462d3ad67721492a24a34df72b8014cf294a34" Dec 05 11:18:04.061884 master-0 kubenswrapper[24928]: I1205 11:18:04.061819 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fjthc"] Dec 05 11:18:04.072823 master-0 kubenswrapper[24928]: I1205 11:18:04.072757 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-fjthc"] Dec 05 11:18:04.218896 master-0 kubenswrapper[24928]: I1205 11:18:04.218832 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b68cc45-2d4a-4377-ba5c-e35ecd4ed578" path="/var/lib/kubelet/pods/7b68cc45-2d4a-4377-ba5c-e35ecd4ed578/volumes" Dec 05 11:18:32.069454 master-0 kubenswrapper[24928]: I1205 11:18:32.068518 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-8tsgs"] Dec 05 11:18:32.084454 master-0 kubenswrapper[24928]: I1205 11:18:32.081689 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tqtr9"] Dec 05 11:18:32.096694 master-0 kubenswrapper[24928]: I1205 11:18:32.096642 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-8tsgs"] Dec 05 11:18:32.110458 master-0 kubenswrapper[24928]: I1205 11:18:32.110357 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-tqtr9"] Dec 05 11:18:32.218073 master-0 kubenswrapper[24928]: I1205 11:18:32.218017 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdfae6e2-22a3-4218-b6e9-2558a61bcead" path="/var/lib/kubelet/pods/cdfae6e2-22a3-4218-b6e9-2558a61bcead/volumes" Dec 05 11:18:32.219104 master-0 kubenswrapper[24928]: I1205 11:18:32.219086 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6dde2d1-2c98-48c1-ac60-02129e7543b3" path="/var/lib/kubelet/pods/f6dde2d1-2c98-48c1-ac60-02129e7543b3/volumes" Dec 05 11:18:48.322819 master-0 kubenswrapper[24928]: I1205 11:18:48.322763 24928 scope.go:117] "RemoveContainer" containerID="5a5890da362f9de00fc5e6f60e48e89a3a543b86628b0d169ed86c857200fe1d" Dec 05 11:18:48.361557 master-0 kubenswrapper[24928]: I1205 11:18:48.361288 24928 scope.go:117] "RemoveContainer" containerID="00fff1acca38683d164302d064fbbf231b93dff900bc630d9a907d5c7c4b4e4d" Dec 05 11:18:48.419087 master-0 kubenswrapper[24928]: I1205 11:18:48.419011 24928 scope.go:117] "RemoveContainer" containerID="fa1dbb07b11ef157ddc386720c998f0bd948fbc44b183a8e3569217c865c86c9" Dec 05 11:18:51.045050 master-0 kubenswrapper[24928]: I1205 11:18:51.044583 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-host-discover-clfhv"] Dec 05 11:18:51.055896 master-0 kubenswrapper[24928]: I1205 11:18:51.055845 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-host-discover-clfhv"] Dec 05 11:18:52.218368 master-0 kubenswrapper[24928]: I1205 11:18:52.218264 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2f41a31-1432-4cb3-a605-b55952b32508" path="/var/lib/kubelet/pods/d2f41a31-1432-4cb3-a605-b55952b32508/volumes" Dec 05 11:18:54.784049 master-0 kubenswrapper[24928]: I1205 11:18:54.783935 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-v6rjc"] Dec 05 11:18:54.793056 master-0 kubenswrapper[24928]: I1205 11:18:54.792972 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-v6rjc"] Dec 05 11:18:56.218083 master-0 kubenswrapper[24928]: I1205 11:18:56.217997 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9837315-d749-43df-b3b0-a7c36eabec7e" path="/var/lib/kubelet/pods/f9837315-d749-43df-b3b0-a7c36eabec7e/volumes" Dec 05 11:19:48.574135 master-0 kubenswrapper[24928]: I1205 11:19:48.574000 24928 scope.go:117] "RemoveContainer" containerID="19d0eb0603192eef0552d5244e0df029fbe0013c6cdb400cfaa38b417c70137a" Dec 05 11:19:48.618581 master-0 kubenswrapper[24928]: I1205 11:19:48.618532 24928 scope.go:117] "RemoveContainer" containerID="e646ca28ebe6fb3f9f871e91271ced75c5814530553a28bcb695d05f7558360c" Dec 05 11:20:37.412708 master-0 kubenswrapper[24928]: I1205 11:20:37.412640 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8c4xh"] Dec 05 11:20:37.413323 master-0 kubenswrapper[24928]: E1205 11:20:37.413161 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b15005f0-2323-46d7-916c-473b2443a139" containerName="registry-server" Dec 05 11:20:37.413323 master-0 kubenswrapper[24928]: I1205 11:20:37.413176 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b15005f0-2323-46d7-916c-473b2443a139" containerName="registry-server" Dec 05 11:20:37.413323 master-0 kubenswrapper[24928]: E1205 11:20:37.413208 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b15005f0-2323-46d7-916c-473b2443a139" containerName="extract-content" Dec 05 11:20:37.413323 master-0 kubenswrapper[24928]: I1205 11:20:37.413214 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b15005f0-2323-46d7-916c-473b2443a139" containerName="extract-content" Dec 05 11:20:37.413323 master-0 kubenswrapper[24928]: E1205 11:20:37.413241 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b15005f0-2323-46d7-916c-473b2443a139" containerName="extract-utilities" Dec 05 11:20:37.413323 master-0 kubenswrapper[24928]: I1205 11:20:37.413249 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="b15005f0-2323-46d7-916c-473b2443a139" containerName="extract-utilities" Dec 05 11:20:37.413563 master-0 kubenswrapper[24928]: I1205 11:20:37.413539 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="b15005f0-2323-46d7-916c-473b2443a139" containerName="registry-server" Dec 05 11:20:37.420369 master-0 kubenswrapper[24928]: I1205 11:20:37.420111 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:37.430701 master-0 kubenswrapper[24928]: I1205 11:20:37.430605 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8c4xh"] Dec 05 11:20:37.475739 master-0 kubenswrapper[24928]: I1205 11:20:37.475480 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-utilities\") pod \"community-operators-8c4xh\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:37.476195 master-0 kubenswrapper[24928]: I1205 11:20:37.476131 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-catalog-content\") pod \"community-operators-8c4xh\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:37.476263 master-0 kubenswrapper[24928]: I1205 11:20:37.476219 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzcmh\" (UniqueName: \"kubernetes.io/projected/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-kube-api-access-pzcmh\") pod \"community-operators-8c4xh\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:37.578296 master-0 kubenswrapper[24928]: I1205 11:20:37.578205 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-utilities\") pod \"community-operators-8c4xh\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:37.578573 master-0 kubenswrapper[24928]: I1205 11:20:37.578408 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-catalog-content\") pod \"community-operators-8c4xh\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:37.578613 master-0 kubenswrapper[24928]: I1205 11:20:37.578581 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzcmh\" (UniqueName: \"kubernetes.io/projected/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-kube-api-access-pzcmh\") pod \"community-operators-8c4xh\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:37.579397 master-0 kubenswrapper[24928]: I1205 11:20:37.579372 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-utilities\") pod \"community-operators-8c4xh\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:37.579667 master-0 kubenswrapper[24928]: I1205 11:20:37.579641 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-catalog-content\") pod \"community-operators-8c4xh\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:37.595157 master-0 kubenswrapper[24928]: I1205 11:20:37.595103 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzcmh\" (UniqueName: \"kubernetes.io/projected/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-kube-api-access-pzcmh\") pod \"community-operators-8c4xh\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:37.749222 master-0 kubenswrapper[24928]: I1205 11:20:37.749151 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:38.336971 master-0 kubenswrapper[24928]: W1205 11:20:38.336911 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod431b5cfe_a8a4_468b_bbb8_3a4e78187a80.slice/crio-cce06329abf78869f183c3c1abce46a96d4574b75115e17e2015863d085079db WatchSource:0}: Error finding container cce06329abf78869f183c3c1abce46a96d4574b75115e17e2015863d085079db: Status 404 returned error can't find the container with id cce06329abf78869f183c3c1abce46a96d4574b75115e17e2015863d085079db Dec 05 11:20:38.354628 master-0 kubenswrapper[24928]: I1205 11:20:38.354531 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8c4xh"] Dec 05 11:20:39.174725 master-0 kubenswrapper[24928]: I1205 11:20:39.174642 24928 generic.go:334] "Generic (PLEG): container finished" podID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" containerID="fc45ca21e1a81bc02e8bda9c48bada759cef40faa7440f711d82246b77070b5d" exitCode=0 Dec 05 11:20:39.174725 master-0 kubenswrapper[24928]: I1205 11:20:39.174715 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8c4xh" event={"ID":"431b5cfe-a8a4-468b-bbb8-3a4e78187a80","Type":"ContainerDied","Data":"fc45ca21e1a81bc02e8bda9c48bada759cef40faa7440f711d82246b77070b5d"} Dec 05 11:20:39.174725 master-0 kubenswrapper[24928]: I1205 11:20:39.174745 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8c4xh" event={"ID":"431b5cfe-a8a4-468b-bbb8-3a4e78187a80","Type":"ContainerStarted","Data":"cce06329abf78869f183c3c1abce46a96d4574b75115e17e2015863d085079db"} Dec 05 11:20:39.177728 master-0 kubenswrapper[24928]: I1205 11:20:39.177664 24928 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:20:40.201398 master-0 kubenswrapper[24928]: I1205 11:20:40.199919 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8c4xh" event={"ID":"431b5cfe-a8a4-468b-bbb8-3a4e78187a80","Type":"ContainerStarted","Data":"2db881ca810a0cd2e643093047d8732f0433b20e3e3722fa4ece294f609080b4"} Dec 05 11:20:41.218316 master-0 kubenswrapper[24928]: I1205 11:20:41.218212 24928 generic.go:334] "Generic (PLEG): container finished" podID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" containerID="2db881ca810a0cd2e643093047d8732f0433b20e3e3722fa4ece294f609080b4" exitCode=0 Dec 05 11:20:41.218911 master-0 kubenswrapper[24928]: I1205 11:20:41.218536 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t7t4q"] Dec 05 11:20:41.221127 master-0 kubenswrapper[24928]: I1205 11:20:41.221078 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8c4xh" event={"ID":"431b5cfe-a8a4-468b-bbb8-3a4e78187a80","Type":"ContainerDied","Data":"2db881ca810a0cd2e643093047d8732f0433b20e3e3722fa4ece294f609080b4"} Dec 05 11:20:41.221212 master-0 kubenswrapper[24928]: I1205 11:20:41.221186 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:41.234652 master-0 kubenswrapper[24928]: I1205 11:20:41.234553 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7t4q"] Dec 05 11:20:41.322598 master-0 kubenswrapper[24928]: I1205 11:20:41.322532 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-catalog-content\") pod \"redhat-marketplace-t7t4q\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:41.324459 master-0 kubenswrapper[24928]: I1205 11:20:41.324383 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-utilities\") pod \"redhat-marketplace-t7t4q\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:41.325986 master-0 kubenswrapper[24928]: I1205 11:20:41.325644 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq7gd\" (UniqueName: \"kubernetes.io/projected/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-kube-api-access-pq7gd\") pod \"redhat-marketplace-t7t4q\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:41.428258 master-0 kubenswrapper[24928]: I1205 11:20:41.428203 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-utilities\") pod \"redhat-marketplace-t7t4q\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:41.428516 master-0 kubenswrapper[24928]: I1205 11:20:41.428323 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq7gd\" (UniqueName: \"kubernetes.io/projected/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-kube-api-access-pq7gd\") pod \"redhat-marketplace-t7t4q\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:41.428516 master-0 kubenswrapper[24928]: I1205 11:20:41.428361 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-catalog-content\") pod \"redhat-marketplace-t7t4q\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:41.428990 master-0 kubenswrapper[24928]: I1205 11:20:41.428799 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-utilities\") pod \"redhat-marketplace-t7t4q\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:41.428990 master-0 kubenswrapper[24928]: I1205 11:20:41.428900 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-catalog-content\") pod \"redhat-marketplace-t7t4q\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:41.447855 master-0 kubenswrapper[24928]: I1205 11:20:41.446846 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq7gd\" (UniqueName: \"kubernetes.io/projected/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-kube-api-access-pq7gd\") pod \"redhat-marketplace-t7t4q\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:41.564368 master-0 kubenswrapper[24928]: I1205 11:20:41.564306 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:42.156604 master-0 kubenswrapper[24928]: I1205 11:20:42.156536 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7t4q"] Dec 05 11:20:42.254310 master-0 kubenswrapper[24928]: I1205 11:20:42.254235 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rxbm6"] Dec 05 11:20:42.257495 master-0 kubenswrapper[24928]: I1205 11:20:42.257451 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8c4xh" event={"ID":"431b5cfe-a8a4-468b-bbb8-3a4e78187a80","Type":"ContainerStarted","Data":"382bf067bc004c32fd69a10340a0dccb1ebddcc47ea1ff03fe4555525b1df629"} Dec 05 11:20:42.257622 master-0 kubenswrapper[24928]: I1205 11:20:42.257499 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rxbm6"] Dec 05 11:20:42.257622 master-0 kubenswrapper[24928]: I1205 11:20:42.257516 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7t4q" event={"ID":"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6","Type":"ContainerStarted","Data":"be75e8b7fed76c294287784f8cf5078842ed08fe1c3654e2fde37bcd22cbd92e"} Dec 05 11:20:42.257752 master-0 kubenswrapper[24928]: I1205 11:20:42.257670 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:42.287203 master-0 kubenswrapper[24928]: I1205 11:20:42.287116 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8c4xh" podStartSLOduration=2.798408057 podStartE2EDuration="5.28709438s" podCreationTimestamp="2025-12-05 11:20:37 +0000 UTC" firstStartedPulling="2025-12-05 11:20:39.177608896 +0000 UTC m=+1999.180802747" lastFinishedPulling="2025-12-05 11:20:41.666295229 +0000 UTC m=+2001.669489070" observedRunningTime="2025-12-05 11:20:42.271491676 +0000 UTC m=+2002.274685537" watchObservedRunningTime="2025-12-05 11:20:42.28709438 +0000 UTC m=+2002.290288231" Dec 05 11:20:42.355668 master-0 kubenswrapper[24928]: I1205 11:20:42.355580 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-utilities\") pod \"redhat-operators-rxbm6\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:42.355933 master-0 kubenswrapper[24928]: I1205 11:20:42.355895 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-catalog-content\") pod \"redhat-operators-rxbm6\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:42.356247 master-0 kubenswrapper[24928]: I1205 11:20:42.356211 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr64p\" (UniqueName: \"kubernetes.io/projected/d7a4621f-3ead-4b91-ad74-bb897868d437-kube-api-access-gr64p\") pod \"redhat-operators-rxbm6\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:42.460944 master-0 kubenswrapper[24928]: I1205 11:20:42.460798 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-catalog-content\") pod \"redhat-operators-rxbm6\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:42.461385 master-0 kubenswrapper[24928]: I1205 11:20:42.461360 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr64p\" (UniqueName: \"kubernetes.io/projected/d7a4621f-3ead-4b91-ad74-bb897868d437-kube-api-access-gr64p\") pod \"redhat-operators-rxbm6\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:42.461671 master-0 kubenswrapper[24928]: I1205 11:20:42.461649 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-utilities\") pod \"redhat-operators-rxbm6\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:42.462505 master-0 kubenswrapper[24928]: I1205 11:20:42.462483 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-utilities\") pod \"redhat-operators-rxbm6\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:42.462951 master-0 kubenswrapper[24928]: I1205 11:20:42.462927 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-catalog-content\") pod \"redhat-operators-rxbm6\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:42.482563 master-0 kubenswrapper[24928]: I1205 11:20:42.482492 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr64p\" (UniqueName: \"kubernetes.io/projected/d7a4621f-3ead-4b91-ad74-bb897868d437-kube-api-access-gr64p\") pod \"redhat-operators-rxbm6\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:42.617367 master-0 kubenswrapper[24928]: I1205 11:20:42.617202 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:43.163533 master-0 kubenswrapper[24928]: W1205 11:20:43.162477 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7a4621f_3ead_4b91_ad74_bb897868d437.slice/crio-a28e6f8d4b0d54f58c97c8b5d46df8b71de865e1d6557e6147b96a8b97f069c6 WatchSource:0}: Error finding container a28e6f8d4b0d54f58c97c8b5d46df8b71de865e1d6557e6147b96a8b97f069c6: Status 404 returned error can't find the container with id a28e6f8d4b0d54f58c97c8b5d46df8b71de865e1d6557e6147b96a8b97f069c6 Dec 05 11:20:43.174686 master-0 kubenswrapper[24928]: I1205 11:20:43.174598 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rxbm6"] Dec 05 11:20:43.280324 master-0 kubenswrapper[24928]: I1205 11:20:43.280276 24928 generic.go:334] "Generic (PLEG): container finished" podID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" containerID="d001fb432a3fb0157ceb5b440a681f896a6b37d49485cb945dd95e0a0f19a766" exitCode=0 Dec 05 11:20:43.287850 master-0 kubenswrapper[24928]: I1205 11:20:43.280887 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7t4q" event={"ID":"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6","Type":"ContainerDied","Data":"d001fb432a3fb0157ceb5b440a681f896a6b37d49485cb945dd95e0a0f19a766"} Dec 05 11:20:43.287850 master-0 kubenswrapper[24928]: I1205 11:20:43.283721 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxbm6" event={"ID":"d7a4621f-3ead-4b91-ad74-bb897868d437","Type":"ContainerStarted","Data":"a28e6f8d4b0d54f58c97c8b5d46df8b71de865e1d6557e6147b96a8b97f069c6"} Dec 05 11:20:44.297931 master-0 kubenswrapper[24928]: I1205 11:20:44.297847 24928 generic.go:334] "Generic (PLEG): container finished" podID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" containerID="cd58d53f21e720dd34a5e70c2098d7267ae23f76c70c18d79415adc07e37cc40" exitCode=0 Dec 05 11:20:44.298617 master-0 kubenswrapper[24928]: I1205 11:20:44.297920 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7t4q" event={"ID":"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6","Type":"ContainerDied","Data":"cd58d53f21e720dd34a5e70c2098d7267ae23f76c70c18d79415adc07e37cc40"} Dec 05 11:20:44.301074 master-0 kubenswrapper[24928]: I1205 11:20:44.301043 24928 generic.go:334] "Generic (PLEG): container finished" podID="d7a4621f-3ead-4b91-ad74-bb897868d437" containerID="411e8f968b4fae497c0814007a815d4e6930487b142273439c2a71a64b9c66d3" exitCode=0 Dec 05 11:20:44.301192 master-0 kubenswrapper[24928]: I1205 11:20:44.301092 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxbm6" event={"ID":"d7a4621f-3ead-4b91-ad74-bb897868d437","Type":"ContainerDied","Data":"411e8f968b4fae497c0814007a815d4e6930487b142273439c2a71a64b9c66d3"} Dec 05 11:20:45.316444 master-0 kubenswrapper[24928]: I1205 11:20:45.316358 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7t4q" event={"ID":"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6","Type":"ContainerStarted","Data":"62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f"} Dec 05 11:20:45.318786 master-0 kubenswrapper[24928]: I1205 11:20:45.318707 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxbm6" event={"ID":"d7a4621f-3ead-4b91-ad74-bb897868d437","Type":"ContainerStarted","Data":"7b2a7a68df6ef2bc8cf723792d30671fdcd31db4a800e7186938168c6b90bbf0"} Dec 05 11:20:45.342835 master-0 kubenswrapper[24928]: I1205 11:20:45.342754 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t7t4q" podStartSLOduration=2.95541694 podStartE2EDuration="4.342737376s" podCreationTimestamp="2025-12-05 11:20:41 +0000 UTC" firstStartedPulling="2025-12-05 11:20:43.282623658 +0000 UTC m=+2003.285817509" lastFinishedPulling="2025-12-05 11:20:44.669944094 +0000 UTC m=+2004.673137945" observedRunningTime="2025-12-05 11:20:45.338622095 +0000 UTC m=+2005.341815946" watchObservedRunningTime="2025-12-05 11:20:45.342737376 +0000 UTC m=+2005.345931227" Dec 05 11:20:46.335867 master-0 kubenswrapper[24928]: I1205 11:20:46.335789 24928 generic.go:334] "Generic (PLEG): container finished" podID="d7a4621f-3ead-4b91-ad74-bb897868d437" containerID="7b2a7a68df6ef2bc8cf723792d30671fdcd31db4a800e7186938168c6b90bbf0" exitCode=0 Dec 05 11:20:46.335867 master-0 kubenswrapper[24928]: I1205 11:20:46.335860 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxbm6" event={"ID":"d7a4621f-3ead-4b91-ad74-bb897868d437","Type":"ContainerDied","Data":"7b2a7a68df6ef2bc8cf723792d30671fdcd31db4a800e7186938168c6b90bbf0"} Dec 05 11:20:47.355480 master-0 kubenswrapper[24928]: I1205 11:20:47.355377 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxbm6" event={"ID":"d7a4621f-3ead-4b91-ad74-bb897868d437","Type":"ContainerStarted","Data":"1c572c67a10bb505a0134d3f31cd6a373e7a89a751c54810789003534b819973"} Dec 05 11:20:47.385242 master-0 kubenswrapper[24928]: I1205 11:20:47.385126 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rxbm6" podStartSLOduration=2.957995673 podStartE2EDuration="5.385106548s" podCreationTimestamp="2025-12-05 11:20:42 +0000 UTC" firstStartedPulling="2025-12-05 11:20:44.303802098 +0000 UTC m=+2004.306995949" lastFinishedPulling="2025-12-05 11:20:46.730912973 +0000 UTC m=+2006.734106824" observedRunningTime="2025-12-05 11:20:47.378875595 +0000 UTC m=+2007.382069446" watchObservedRunningTime="2025-12-05 11:20:47.385106548 +0000 UTC m=+2007.388300389" Dec 05 11:20:47.750216 master-0 kubenswrapper[24928]: I1205 11:20:47.750086 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:47.750216 master-0 kubenswrapper[24928]: I1205 11:20:47.750141 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:47.803802 master-0 kubenswrapper[24928]: I1205 11:20:47.803720 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:48.455308 master-0 kubenswrapper[24928]: I1205 11:20:48.455215 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:51.088971 master-0 kubenswrapper[24928]: I1205 11:20:51.088884 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8c4xh"] Dec 05 11:20:51.089616 master-0 kubenswrapper[24928]: I1205 11:20:51.089171 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8c4xh" podUID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" containerName="registry-server" containerID="cri-o://382bf067bc004c32fd69a10340a0dccb1ebddcc47ea1ff03fe4555525b1df629" gracePeriod=2 Dec 05 11:20:51.412378 master-0 kubenswrapper[24928]: I1205 11:20:51.412194 24928 generic.go:334] "Generic (PLEG): container finished" podID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" containerID="382bf067bc004c32fd69a10340a0dccb1ebddcc47ea1ff03fe4555525b1df629" exitCode=0 Dec 05 11:20:51.412378 master-0 kubenswrapper[24928]: I1205 11:20:51.412310 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8c4xh" event={"ID":"431b5cfe-a8a4-468b-bbb8-3a4e78187a80","Type":"ContainerDied","Data":"382bf067bc004c32fd69a10340a0dccb1ebddcc47ea1ff03fe4555525b1df629"} Dec 05 11:20:51.565233 master-0 kubenswrapper[24928]: I1205 11:20:51.565167 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:51.565233 master-0 kubenswrapper[24928]: I1205 11:20:51.565227 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:51.622735 master-0 kubenswrapper[24928]: I1205 11:20:51.622673 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:51.962053 master-0 kubenswrapper[24928]: I1205 11:20:51.961981 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:52.046984 master-0 kubenswrapper[24928]: I1205 11:20:52.046875 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-catalog-content\") pod \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " Dec 05 11:20:52.047231 master-0 kubenswrapper[24928]: I1205 11:20:52.047050 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzcmh\" (UniqueName: \"kubernetes.io/projected/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-kube-api-access-pzcmh\") pod \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " Dec 05 11:20:52.047231 master-0 kubenswrapper[24928]: I1205 11:20:52.047132 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-utilities\") pod \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\" (UID: \"431b5cfe-a8a4-468b-bbb8-3a4e78187a80\") " Dec 05 11:20:52.048934 master-0 kubenswrapper[24928]: I1205 11:20:52.048890 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-utilities" (OuterVolumeSpecName: "utilities") pod "431b5cfe-a8a4-468b-bbb8-3a4e78187a80" (UID: "431b5cfe-a8a4-468b-bbb8-3a4e78187a80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:20:52.050574 master-0 kubenswrapper[24928]: I1205 11:20:52.050520 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-kube-api-access-pzcmh" (OuterVolumeSpecName: "kube-api-access-pzcmh") pod "431b5cfe-a8a4-468b-bbb8-3a4e78187a80" (UID: "431b5cfe-a8a4-468b-bbb8-3a4e78187a80"). InnerVolumeSpecName "kube-api-access-pzcmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:20:52.106140 master-0 kubenswrapper[24928]: I1205 11:20:52.106093 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "431b5cfe-a8a4-468b-bbb8-3a4e78187a80" (UID: "431b5cfe-a8a4-468b-bbb8-3a4e78187a80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:20:52.152126 master-0 kubenswrapper[24928]: I1205 11:20:52.151956 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:20:52.152126 master-0 kubenswrapper[24928]: I1205 11:20:52.151999 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzcmh\" (UniqueName: \"kubernetes.io/projected/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-kube-api-access-pzcmh\") on node \"master-0\" DevicePath \"\"" Dec 05 11:20:52.152126 master-0 kubenswrapper[24928]: I1205 11:20:52.152009 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/431b5cfe-a8a4-468b-bbb8-3a4e78187a80-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:20:52.424106 master-0 kubenswrapper[24928]: I1205 11:20:52.423938 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8c4xh" event={"ID":"431b5cfe-a8a4-468b-bbb8-3a4e78187a80","Type":"ContainerDied","Data":"cce06329abf78869f183c3c1abce46a96d4574b75115e17e2015863d085079db"} Dec 05 11:20:52.424106 master-0 kubenswrapper[24928]: I1205 11:20:52.423977 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8c4xh" Dec 05 11:20:52.424106 master-0 kubenswrapper[24928]: I1205 11:20:52.424016 24928 scope.go:117] "RemoveContainer" containerID="382bf067bc004c32fd69a10340a0dccb1ebddcc47ea1ff03fe4555525b1df629" Dec 05 11:20:52.468800 master-0 kubenswrapper[24928]: I1205 11:20:52.468708 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8c4xh"] Dec 05 11:20:52.472458 master-0 kubenswrapper[24928]: I1205 11:20:52.472399 24928 scope.go:117] "RemoveContainer" containerID="2db881ca810a0cd2e643093047d8732f0433b20e3e3722fa4ece294f609080b4" Dec 05 11:20:52.479529 master-0 kubenswrapper[24928]: I1205 11:20:52.479477 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8c4xh"] Dec 05 11:20:52.495341 master-0 kubenswrapper[24928]: I1205 11:20:52.495260 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:52.501387 master-0 kubenswrapper[24928]: I1205 11:20:52.501307 24928 scope.go:117] "RemoveContainer" containerID="fc45ca21e1a81bc02e8bda9c48bada759cef40faa7440f711d82246b77070b5d" Dec 05 11:20:52.618350 master-0 kubenswrapper[24928]: I1205 11:20:52.618241 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:52.618673 master-0 kubenswrapper[24928]: I1205 11:20:52.618354 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:52.679597 master-0 kubenswrapper[24928]: I1205 11:20:52.679362 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:53.484834 master-0 kubenswrapper[24928]: I1205 11:20:53.484789 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:54.194530 master-0 kubenswrapper[24928]: I1205 11:20:54.194341 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7t4q"] Dec 05 11:20:54.219753 master-0 kubenswrapper[24928]: I1205 11:20:54.219671 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" path="/var/lib/kubelet/pods/431b5cfe-a8a4-468b-bbb8-3a4e78187a80/volumes" Dec 05 11:20:54.448980 master-0 kubenswrapper[24928]: I1205 11:20:54.448768 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t7t4q" podUID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" containerName="registry-server" containerID="cri-o://62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f" gracePeriod=2 Dec 05 11:20:55.014442 master-0 kubenswrapper[24928]: I1205 11:20:55.014366 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:55.128336 master-0 kubenswrapper[24928]: I1205 11:20:55.128269 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-catalog-content\") pod \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " Dec 05 11:20:55.128566 master-0 kubenswrapper[24928]: I1205 11:20:55.128418 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-utilities\") pod \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " Dec 05 11:20:55.128640 master-0 kubenswrapper[24928]: I1205 11:20:55.128613 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq7gd\" (UniqueName: \"kubernetes.io/projected/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-kube-api-access-pq7gd\") pod \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\" (UID: \"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6\") " Dec 05 11:20:55.129208 master-0 kubenswrapper[24928]: I1205 11:20:55.129170 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-utilities" (OuterVolumeSpecName: "utilities") pod "9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" (UID: "9036c88b-237d-4ef8-8ec8-c4499c2c6dc6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:20:55.129326 master-0 kubenswrapper[24928]: I1205 11:20:55.129296 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:20:55.132234 master-0 kubenswrapper[24928]: I1205 11:20:55.132204 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-kube-api-access-pq7gd" (OuterVolumeSpecName: "kube-api-access-pq7gd") pod "9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" (UID: "9036c88b-237d-4ef8-8ec8-c4499c2c6dc6"). InnerVolumeSpecName "kube-api-access-pq7gd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:20:55.146525 master-0 kubenswrapper[24928]: I1205 11:20:55.146443 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" (UID: "9036c88b-237d-4ef8-8ec8-c4499c2c6dc6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:20:55.232805 master-0 kubenswrapper[24928]: I1205 11:20:55.232719 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq7gd\" (UniqueName: \"kubernetes.io/projected/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-kube-api-access-pq7gd\") on node \"master-0\" DevicePath \"\"" Dec 05 11:20:55.233077 master-0 kubenswrapper[24928]: I1205 11:20:55.232835 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:20:55.474562 master-0 kubenswrapper[24928]: I1205 11:20:55.474385 24928 generic.go:334] "Generic (PLEG): container finished" podID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" containerID="62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f" exitCode=0 Dec 05 11:20:55.474562 master-0 kubenswrapper[24928]: I1205 11:20:55.474458 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7t4q" event={"ID":"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6","Type":"ContainerDied","Data":"62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f"} Dec 05 11:20:55.474562 master-0 kubenswrapper[24928]: I1205 11:20:55.474522 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t7t4q" event={"ID":"9036c88b-237d-4ef8-8ec8-c4499c2c6dc6","Type":"ContainerDied","Data":"be75e8b7fed76c294287784f8cf5078842ed08fe1c3654e2fde37bcd22cbd92e"} Dec 05 11:20:55.474562 master-0 kubenswrapper[24928]: I1205 11:20:55.474543 24928 scope.go:117] "RemoveContainer" containerID="62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f" Dec 05 11:20:55.474916 master-0 kubenswrapper[24928]: I1205 11:20:55.474656 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t7t4q" Dec 05 11:20:55.501785 master-0 kubenswrapper[24928]: I1205 11:20:55.501488 24928 scope.go:117] "RemoveContainer" containerID="cd58d53f21e720dd34a5e70c2098d7267ae23f76c70c18d79415adc07e37cc40" Dec 05 11:20:55.531930 master-0 kubenswrapper[24928]: I1205 11:20:55.531877 24928 scope.go:117] "RemoveContainer" containerID="d001fb432a3fb0157ceb5b440a681f896a6b37d49485cb945dd95e0a0f19a766" Dec 05 11:20:55.538658 master-0 kubenswrapper[24928]: I1205 11:20:55.538574 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7t4q"] Dec 05 11:20:55.557246 master-0 kubenswrapper[24928]: I1205 11:20:55.557110 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t7t4q"] Dec 05 11:20:55.595678 master-0 kubenswrapper[24928]: I1205 11:20:55.595606 24928 scope.go:117] "RemoveContainer" containerID="62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f" Dec 05 11:20:55.596107 master-0 kubenswrapper[24928]: E1205 11:20:55.596053 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f\": container with ID starting with 62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f not found: ID does not exist" containerID="62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f" Dec 05 11:20:55.596161 master-0 kubenswrapper[24928]: I1205 11:20:55.596112 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f"} err="failed to get container status \"62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f\": rpc error: code = NotFound desc = could not find container \"62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f\": container with ID starting with 62d0c85eb1617fc4f8a1457d97e3ac7236563ddfc17c8cbc9bb448e71c640c4f not found: ID does not exist" Dec 05 11:20:55.596161 master-0 kubenswrapper[24928]: I1205 11:20:55.596147 24928 scope.go:117] "RemoveContainer" containerID="cd58d53f21e720dd34a5e70c2098d7267ae23f76c70c18d79415adc07e37cc40" Dec 05 11:20:55.596652 master-0 kubenswrapper[24928]: E1205 11:20:55.596608 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd58d53f21e720dd34a5e70c2098d7267ae23f76c70c18d79415adc07e37cc40\": container with ID starting with cd58d53f21e720dd34a5e70c2098d7267ae23f76c70c18d79415adc07e37cc40 not found: ID does not exist" containerID="cd58d53f21e720dd34a5e70c2098d7267ae23f76c70c18d79415adc07e37cc40" Dec 05 11:20:55.596712 master-0 kubenswrapper[24928]: I1205 11:20:55.596653 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd58d53f21e720dd34a5e70c2098d7267ae23f76c70c18d79415adc07e37cc40"} err="failed to get container status \"cd58d53f21e720dd34a5e70c2098d7267ae23f76c70c18d79415adc07e37cc40\": rpc error: code = NotFound desc = could not find container \"cd58d53f21e720dd34a5e70c2098d7267ae23f76c70c18d79415adc07e37cc40\": container with ID starting with cd58d53f21e720dd34a5e70c2098d7267ae23f76c70c18d79415adc07e37cc40 not found: ID does not exist" Dec 05 11:20:55.596712 master-0 kubenswrapper[24928]: I1205 11:20:55.596686 24928 scope.go:117] "RemoveContainer" containerID="d001fb432a3fb0157ceb5b440a681f896a6b37d49485cb945dd95e0a0f19a766" Dec 05 11:20:55.597390 master-0 kubenswrapper[24928]: E1205 11:20:55.597350 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d001fb432a3fb0157ceb5b440a681f896a6b37d49485cb945dd95e0a0f19a766\": container with ID starting with d001fb432a3fb0157ceb5b440a681f896a6b37d49485cb945dd95e0a0f19a766 not found: ID does not exist" containerID="d001fb432a3fb0157ceb5b440a681f896a6b37d49485cb945dd95e0a0f19a766" Dec 05 11:20:55.597470 master-0 kubenswrapper[24928]: I1205 11:20:55.597385 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d001fb432a3fb0157ceb5b440a681f896a6b37d49485cb945dd95e0a0f19a766"} err="failed to get container status \"d001fb432a3fb0157ceb5b440a681f896a6b37d49485cb945dd95e0a0f19a766\": rpc error: code = NotFound desc = could not find container \"d001fb432a3fb0157ceb5b440a681f896a6b37d49485cb945dd95e0a0f19a766\": container with ID starting with d001fb432a3fb0157ceb5b440a681f896a6b37d49485cb945dd95e0a0f19a766 not found: ID does not exist" Dec 05 11:20:56.247347 master-0 kubenswrapper[24928]: I1205 11:20:56.247261 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" path="/var/lib/kubelet/pods/9036c88b-237d-4ef8-8ec8-c4499c2c6dc6/volumes" Dec 05 11:20:57.305790 master-0 kubenswrapper[24928]: I1205 11:20:57.304227 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rxbm6"] Dec 05 11:20:57.306903 master-0 kubenswrapper[24928]: I1205 11:20:57.306822 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rxbm6" podUID="d7a4621f-3ead-4b91-ad74-bb897868d437" containerName="registry-server" containerID="cri-o://1c572c67a10bb505a0134d3f31cd6a373e7a89a751c54810789003534b819973" gracePeriod=2 Dec 05 11:20:58.512306 master-0 kubenswrapper[24928]: I1205 11:20:58.512230 24928 generic.go:334] "Generic (PLEG): container finished" podID="d7a4621f-3ead-4b91-ad74-bb897868d437" containerID="1c572c67a10bb505a0134d3f31cd6a373e7a89a751c54810789003534b819973" exitCode=0 Dec 05 11:20:58.512306 master-0 kubenswrapper[24928]: I1205 11:20:58.512296 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxbm6" event={"ID":"d7a4621f-3ead-4b91-ad74-bb897868d437","Type":"ContainerDied","Data":"1c572c67a10bb505a0134d3f31cd6a373e7a89a751c54810789003534b819973"} Dec 05 11:20:58.512932 master-0 kubenswrapper[24928]: I1205 11:20:58.512331 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rxbm6" event={"ID":"d7a4621f-3ead-4b91-ad74-bb897868d437","Type":"ContainerDied","Data":"a28e6f8d4b0d54f58c97c8b5d46df8b71de865e1d6557e6147b96a8b97f069c6"} Dec 05 11:20:58.512932 master-0 kubenswrapper[24928]: I1205 11:20:58.512348 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a28e6f8d4b0d54f58c97c8b5d46df8b71de865e1d6557e6147b96a8b97f069c6" Dec 05 11:20:58.576737 master-0 kubenswrapper[24928]: I1205 11:20:58.576667 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:58.623655 master-0 kubenswrapper[24928]: I1205 11:20:58.620709 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-catalog-content\") pod \"d7a4621f-3ead-4b91-ad74-bb897868d437\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " Dec 05 11:20:58.623655 master-0 kubenswrapper[24928]: I1205 11:20:58.620921 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gr64p\" (UniqueName: \"kubernetes.io/projected/d7a4621f-3ead-4b91-ad74-bb897868d437-kube-api-access-gr64p\") pod \"d7a4621f-3ead-4b91-ad74-bb897868d437\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " Dec 05 11:20:58.623655 master-0 kubenswrapper[24928]: I1205 11:20:58.621062 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-utilities\") pod \"d7a4621f-3ead-4b91-ad74-bb897868d437\" (UID: \"d7a4621f-3ead-4b91-ad74-bb897868d437\") " Dec 05 11:20:58.623655 master-0 kubenswrapper[24928]: I1205 11:20:58.622095 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-utilities" (OuterVolumeSpecName: "utilities") pod "d7a4621f-3ead-4b91-ad74-bb897868d437" (UID: "d7a4621f-3ead-4b91-ad74-bb897868d437"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:20:58.624040 master-0 kubenswrapper[24928]: I1205 11:20:58.623932 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7a4621f-3ead-4b91-ad74-bb897868d437-kube-api-access-gr64p" (OuterVolumeSpecName: "kube-api-access-gr64p") pod "d7a4621f-3ead-4b91-ad74-bb897868d437" (UID: "d7a4621f-3ead-4b91-ad74-bb897868d437"). InnerVolumeSpecName "kube-api-access-gr64p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:20:58.725108 master-0 kubenswrapper[24928]: I1205 11:20:58.724899 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:20:58.725355 master-0 kubenswrapper[24928]: I1205 11:20:58.725172 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gr64p\" (UniqueName: \"kubernetes.io/projected/d7a4621f-3ead-4b91-ad74-bb897868d437-kube-api-access-gr64p\") on node \"master-0\" DevicePath \"\"" Dec 05 11:20:58.751100 master-0 kubenswrapper[24928]: I1205 11:20:58.751029 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7a4621f-3ead-4b91-ad74-bb897868d437" (UID: "d7a4621f-3ead-4b91-ad74-bb897868d437"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:20:58.827677 master-0 kubenswrapper[24928]: I1205 11:20:58.827586 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7a4621f-3ead-4b91-ad74-bb897868d437-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:20:59.522732 master-0 kubenswrapper[24928]: I1205 11:20:59.522623 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rxbm6" Dec 05 11:20:59.570576 master-0 kubenswrapper[24928]: I1205 11:20:59.570498 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rxbm6"] Dec 05 11:20:59.581463 master-0 kubenswrapper[24928]: I1205 11:20:59.581381 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rxbm6"] Dec 05 11:21:00.218260 master-0 kubenswrapper[24928]: I1205 11:21:00.218019 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7a4621f-3ead-4b91-ad74-bb897868d437" path="/var/lib/kubelet/pods/d7a4621f-3ead-4b91-ad74-bb897868d437/volumes" Dec 05 11:25:54.605671 master-0 kubenswrapper[24928]: I1205 11:25:54.605580 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s7dnk"] Dec 05 11:25:54.607502 master-0 kubenswrapper[24928]: E1205 11:25:54.607406 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" containerName="extract-utilities" Dec 05 11:25:54.607644 master-0 kubenswrapper[24928]: I1205 11:25:54.607519 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" containerName="extract-utilities" Dec 05 11:25:54.607719 master-0 kubenswrapper[24928]: E1205 11:25:54.607671 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" containerName="registry-server" Dec 05 11:25:54.607719 master-0 kubenswrapper[24928]: I1205 11:25:54.607684 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" containerName="registry-server" Dec 05 11:25:54.607719 master-0 kubenswrapper[24928]: E1205 11:25:54.607700 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" containerName="extract-content" Dec 05 11:25:54.607719 master-0 kubenswrapper[24928]: I1205 11:25:54.607714 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" containerName="extract-content" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: E1205 11:25:54.607731 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a4621f-3ead-4b91-ad74-bb897868d437" containerName="extract-utilities" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: I1205 11:25:54.607739 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a4621f-3ead-4b91-ad74-bb897868d437" containerName="extract-utilities" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: E1205 11:25:54.607752 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" containerName="extract-content" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: I1205 11:25:54.607761 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" containerName="extract-content" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: E1205 11:25:54.607779 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a4621f-3ead-4b91-ad74-bb897868d437" containerName="extract-content" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: I1205 11:25:54.607788 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a4621f-3ead-4b91-ad74-bb897868d437" containerName="extract-content" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: E1205 11:25:54.607803 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7a4621f-3ead-4b91-ad74-bb897868d437" containerName="registry-server" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: I1205 11:25:54.607811 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7a4621f-3ead-4b91-ad74-bb897868d437" containerName="registry-server" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: E1205 11:25:54.607836 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" containerName="registry-server" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: I1205 11:25:54.607848 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" containerName="registry-server" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: E1205 11:25:54.607869 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" containerName="extract-utilities" Dec 05 11:25:54.607920 master-0 kubenswrapper[24928]: I1205 11:25:54.607877 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" containerName="extract-utilities" Dec 05 11:25:54.608447 master-0 kubenswrapper[24928]: I1205 11:25:54.608182 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="9036c88b-237d-4ef8-8ec8-c4499c2c6dc6" containerName="registry-server" Dec 05 11:25:54.608447 master-0 kubenswrapper[24928]: I1205 11:25:54.608213 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="431b5cfe-a8a4-468b-bbb8-3a4e78187a80" containerName="registry-server" Dec 05 11:25:54.608447 master-0 kubenswrapper[24928]: I1205 11:25:54.608229 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7a4621f-3ead-4b91-ad74-bb897868d437" containerName="registry-server" Dec 05 11:25:54.610170 master-0 kubenswrapper[24928]: I1205 11:25:54.610128 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:25:54.622197 master-0 kubenswrapper[24928]: I1205 11:25:54.622123 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s7dnk"] Dec 05 11:25:54.758968 master-0 kubenswrapper[24928]: I1205 11:25:54.758903 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-utilities\") pod \"certified-operators-s7dnk\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:25:54.759547 master-0 kubenswrapper[24928]: I1205 11:25:54.759407 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gpkl\" (UniqueName: \"kubernetes.io/projected/acbbbd49-2b07-4978-8462-dc8ce16d82b0-kube-api-access-6gpkl\") pod \"certified-operators-s7dnk\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:25:54.759811 master-0 kubenswrapper[24928]: I1205 11:25:54.759789 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-catalog-content\") pod \"certified-operators-s7dnk\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:25:54.863140 master-0 kubenswrapper[24928]: I1205 11:25:54.862985 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gpkl\" (UniqueName: \"kubernetes.io/projected/acbbbd49-2b07-4978-8462-dc8ce16d82b0-kube-api-access-6gpkl\") pod \"certified-operators-s7dnk\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:25:54.863140 master-0 kubenswrapper[24928]: I1205 11:25:54.863148 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-catalog-content\") pod \"certified-operators-s7dnk\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:25:54.863436 master-0 kubenswrapper[24928]: I1205 11:25:54.863259 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-utilities\") pod \"certified-operators-s7dnk\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:25:54.863966 master-0 kubenswrapper[24928]: I1205 11:25:54.863934 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-utilities\") pod \"certified-operators-s7dnk\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:25:54.864035 master-0 kubenswrapper[24928]: I1205 11:25:54.863940 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-catalog-content\") pod \"certified-operators-s7dnk\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:25:55.019879 master-0 kubenswrapper[24928]: I1205 11:25:55.019836 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gpkl\" (UniqueName: \"kubernetes.io/projected/acbbbd49-2b07-4978-8462-dc8ce16d82b0-kube-api-access-6gpkl\") pod \"certified-operators-s7dnk\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:25:55.237142 master-0 kubenswrapper[24928]: I1205 11:25:55.236553 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:25:55.771513 master-0 kubenswrapper[24928]: W1205 11:25:55.768682 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podacbbbd49_2b07_4978_8462_dc8ce16d82b0.slice/crio-4d4704645189bb9d9cc530232a5c3f6bf5024ce4452fc71fac9bd5906bed796c WatchSource:0}: Error finding container 4d4704645189bb9d9cc530232a5c3f6bf5024ce4452fc71fac9bd5906bed796c: Status 404 returned error can't find the container with id 4d4704645189bb9d9cc530232a5c3f6bf5024ce4452fc71fac9bd5906bed796c Dec 05 11:25:55.772870 master-0 kubenswrapper[24928]: I1205 11:25:55.772790 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s7dnk"] Dec 05 11:25:56.315280 master-0 kubenswrapper[24928]: I1205 11:25:56.315225 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s7dnk" event={"ID":"acbbbd49-2b07-4978-8462-dc8ce16d82b0","Type":"ContainerStarted","Data":"6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a"} Dec 05 11:25:56.315565 master-0 kubenswrapper[24928]: I1205 11:25:56.315549 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s7dnk" event={"ID":"acbbbd49-2b07-4978-8462-dc8ce16d82b0","Type":"ContainerStarted","Data":"4d4704645189bb9d9cc530232a5c3f6bf5024ce4452fc71fac9bd5906bed796c"} Dec 05 11:25:57.328446 master-0 kubenswrapper[24928]: I1205 11:25:57.328352 24928 generic.go:334] "Generic (PLEG): container finished" podID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" containerID="6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a" exitCode=0 Dec 05 11:25:57.328999 master-0 kubenswrapper[24928]: I1205 11:25:57.328494 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s7dnk" event={"ID":"acbbbd49-2b07-4978-8462-dc8ce16d82b0","Type":"ContainerDied","Data":"6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a"} Dec 05 11:25:57.330319 master-0 kubenswrapper[24928]: I1205 11:25:57.330284 24928 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:25:58.340559 master-0 kubenswrapper[24928]: I1205 11:25:58.340485 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s7dnk" event={"ID":"acbbbd49-2b07-4978-8462-dc8ce16d82b0","Type":"ContainerStarted","Data":"cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1"} Dec 05 11:25:59.359602 master-0 kubenswrapper[24928]: I1205 11:25:59.359530 24928 generic.go:334] "Generic (PLEG): container finished" podID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" containerID="cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1" exitCode=0 Dec 05 11:25:59.360237 master-0 kubenswrapper[24928]: I1205 11:25:59.359623 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s7dnk" event={"ID":"acbbbd49-2b07-4978-8462-dc8ce16d82b0","Type":"ContainerDied","Data":"cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1"} Dec 05 11:26:01.384595 master-0 kubenswrapper[24928]: I1205 11:26:01.384529 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s7dnk" event={"ID":"acbbbd49-2b07-4978-8462-dc8ce16d82b0","Type":"ContainerStarted","Data":"774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18"} Dec 05 11:26:01.807849 master-0 kubenswrapper[24928]: I1205 11:26:01.807709 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s7dnk" podStartSLOduration=4.403277729 podStartE2EDuration="7.807668042s" podCreationTimestamp="2025-12-05 11:25:54 +0000 UTC" firstStartedPulling="2025-12-05 11:25:57.330245311 +0000 UTC m=+2317.333439162" lastFinishedPulling="2025-12-05 11:26:00.734635604 +0000 UTC m=+2320.737829475" observedRunningTime="2025-12-05 11:26:01.801400587 +0000 UTC m=+2321.804594438" watchObservedRunningTime="2025-12-05 11:26:01.807668042 +0000 UTC m=+2321.810861893" Dec 05 11:26:05.238115 master-0 kubenswrapper[24928]: I1205 11:26:05.238029 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:26:05.238115 master-0 kubenswrapper[24928]: I1205 11:26:05.238094 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:26:05.299134 master-0 kubenswrapper[24928]: I1205 11:26:05.299070 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:26:05.503594 master-0 kubenswrapper[24928]: I1205 11:26:05.503428 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:26:05.564075 master-0 kubenswrapper[24928]: I1205 11:26:05.563984 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s7dnk"] Dec 05 11:26:07.481842 master-0 kubenswrapper[24928]: I1205 11:26:07.481770 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s7dnk" podUID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" containerName="registry-server" containerID="cri-o://774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18" gracePeriod=2 Dec 05 11:26:08.175234 master-0 kubenswrapper[24928]: I1205 11:26:08.175167 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:26:08.335034 master-0 kubenswrapper[24928]: I1205 11:26:08.334948 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-utilities\") pod \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " Dec 05 11:26:08.335354 master-0 kubenswrapper[24928]: I1205 11:26:08.335068 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gpkl\" (UniqueName: \"kubernetes.io/projected/acbbbd49-2b07-4978-8462-dc8ce16d82b0-kube-api-access-6gpkl\") pod \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " Dec 05 11:26:08.335354 master-0 kubenswrapper[24928]: I1205 11:26:08.335148 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-catalog-content\") pod \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\" (UID: \"acbbbd49-2b07-4978-8462-dc8ce16d82b0\") " Dec 05 11:26:08.336801 master-0 kubenswrapper[24928]: I1205 11:26:08.336742 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-utilities" (OuterVolumeSpecName: "utilities") pod "acbbbd49-2b07-4978-8462-dc8ce16d82b0" (UID: "acbbbd49-2b07-4978-8462-dc8ce16d82b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:26:08.349064 master-0 kubenswrapper[24928]: I1205 11:26:08.348985 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acbbbd49-2b07-4978-8462-dc8ce16d82b0-kube-api-access-6gpkl" (OuterVolumeSpecName: "kube-api-access-6gpkl") pod "acbbbd49-2b07-4978-8462-dc8ce16d82b0" (UID: "acbbbd49-2b07-4978-8462-dc8ce16d82b0"). InnerVolumeSpecName "kube-api-access-6gpkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:26:08.436272 master-0 kubenswrapper[24928]: I1205 11:26:08.436163 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "acbbbd49-2b07-4978-8462-dc8ce16d82b0" (UID: "acbbbd49-2b07-4978-8462-dc8ce16d82b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:26:08.438608 master-0 kubenswrapper[24928]: I1205 11:26:08.438551 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:26:08.438608 master-0 kubenswrapper[24928]: I1205 11:26:08.438602 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/acbbbd49-2b07-4978-8462-dc8ce16d82b0-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:26:08.438762 master-0 kubenswrapper[24928]: I1205 11:26:08.438618 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gpkl\" (UniqueName: \"kubernetes.io/projected/acbbbd49-2b07-4978-8462-dc8ce16d82b0-kube-api-access-6gpkl\") on node \"master-0\" DevicePath \"\"" Dec 05 11:26:08.574459 master-0 kubenswrapper[24928]: I1205 11:26:08.571771 24928 generic.go:334] "Generic (PLEG): container finished" podID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" containerID="774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18" exitCode=0 Dec 05 11:26:08.574459 master-0 kubenswrapper[24928]: I1205 11:26:08.571855 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s7dnk" event={"ID":"acbbbd49-2b07-4978-8462-dc8ce16d82b0","Type":"ContainerDied","Data":"774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18"} Dec 05 11:26:08.574459 master-0 kubenswrapper[24928]: I1205 11:26:08.571891 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s7dnk" event={"ID":"acbbbd49-2b07-4978-8462-dc8ce16d82b0","Type":"ContainerDied","Data":"4d4704645189bb9d9cc530232a5c3f6bf5024ce4452fc71fac9bd5906bed796c"} Dec 05 11:26:08.574459 master-0 kubenswrapper[24928]: I1205 11:26:08.571915 24928 scope.go:117] "RemoveContainer" containerID="774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18" Dec 05 11:26:08.574459 master-0 kubenswrapper[24928]: I1205 11:26:08.572107 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s7dnk" Dec 05 11:26:08.621052 master-0 kubenswrapper[24928]: I1205 11:26:08.619146 24928 scope.go:117] "RemoveContainer" containerID="cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1" Dec 05 11:26:08.640582 master-0 kubenswrapper[24928]: I1205 11:26:08.640450 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s7dnk"] Dec 05 11:26:08.650293 master-0 kubenswrapper[24928]: I1205 11:26:08.650240 24928 scope.go:117] "RemoveContainer" containerID="6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a" Dec 05 11:26:08.656325 master-0 kubenswrapper[24928]: I1205 11:26:08.656264 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s7dnk"] Dec 05 11:26:08.696718 master-0 kubenswrapper[24928]: I1205 11:26:08.696658 24928 scope.go:117] "RemoveContainer" containerID="774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18" Dec 05 11:26:08.697235 master-0 kubenswrapper[24928]: E1205 11:26:08.697182 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18\": container with ID starting with 774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18 not found: ID does not exist" containerID="774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18" Dec 05 11:26:08.697306 master-0 kubenswrapper[24928]: I1205 11:26:08.697238 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18"} err="failed to get container status \"774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18\": rpc error: code = NotFound desc = could not find container \"774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18\": container with ID starting with 774b80e7d0d98c0c326184409009a80bf696dbdf9fa855e093a496f63f12db18 not found: ID does not exist" Dec 05 11:26:08.697306 master-0 kubenswrapper[24928]: I1205 11:26:08.697276 24928 scope.go:117] "RemoveContainer" containerID="cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1" Dec 05 11:26:08.697731 master-0 kubenswrapper[24928]: E1205 11:26:08.697675 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1\": container with ID starting with cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1 not found: ID does not exist" containerID="cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1" Dec 05 11:26:08.697801 master-0 kubenswrapper[24928]: I1205 11:26:08.697733 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1"} err="failed to get container status \"cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1\": rpc error: code = NotFound desc = could not find container \"cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1\": container with ID starting with cb0de079e2b99cae3d84b0c8665e0d7a67e077da5ff8eeda8c4514a3209855e1 not found: ID does not exist" Dec 05 11:26:08.697801 master-0 kubenswrapper[24928]: I1205 11:26:08.697762 24928 scope.go:117] "RemoveContainer" containerID="6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a" Dec 05 11:26:08.698139 master-0 kubenswrapper[24928]: E1205 11:26:08.698092 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a\": container with ID starting with 6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a not found: ID does not exist" containerID="6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a" Dec 05 11:26:08.698212 master-0 kubenswrapper[24928]: I1205 11:26:08.698141 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a"} err="failed to get container status \"6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a\": rpc error: code = NotFound desc = could not find container \"6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a\": container with ID starting with 6195fe41b5d1813ebb633cf258e1cab1a5643d495194916cedb50d54b0e0d68a not found: ID does not exist" Dec 05 11:26:10.226705 master-0 kubenswrapper[24928]: I1205 11:26:10.226596 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" path="/var/lib/kubelet/pods/acbbbd49-2b07-4978-8462-dc8ce16d82b0/volumes" Dec 05 11:26:48.952543 master-0 kubenswrapper[24928]: I1205 11:26:48.952480 24928 scope.go:117] "RemoveContainer" containerID="411e8f968b4fae497c0814007a815d4e6930487b142273439c2a71a64b9c66d3" Dec 05 11:26:48.995281 master-0 kubenswrapper[24928]: I1205 11:26:48.995172 24928 scope.go:117] "RemoveContainer" containerID="7b2a7a68df6ef2bc8cf723792d30671fdcd31db4a800e7186938168c6b90bbf0" Dec 05 11:26:49.053638 master-0 kubenswrapper[24928]: I1205 11:26:49.053584 24928 scope.go:117] "RemoveContainer" containerID="1c572c67a10bb505a0134d3f31cd6a373e7a89a751c54810789003534b819973" Dec 05 11:30:01.165557 master-0 kubenswrapper[24928]: I1205 11:30:01.165483 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv"] Dec 05 11:30:01.166390 master-0 kubenswrapper[24928]: E1205 11:30:01.166135 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" containerName="extract-utilities" Dec 05 11:30:01.166390 master-0 kubenswrapper[24928]: I1205 11:30:01.166151 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" containerName="extract-utilities" Dec 05 11:30:01.166390 master-0 kubenswrapper[24928]: E1205 11:30:01.166176 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" containerName="extract-content" Dec 05 11:30:01.166390 master-0 kubenswrapper[24928]: I1205 11:30:01.166183 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" containerName="extract-content" Dec 05 11:30:01.166390 master-0 kubenswrapper[24928]: E1205 11:30:01.166205 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" containerName="registry-server" Dec 05 11:30:01.166390 master-0 kubenswrapper[24928]: I1205 11:30:01.166213 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" containerName="registry-server" Dec 05 11:30:01.166794 master-0 kubenswrapper[24928]: I1205 11:30:01.166533 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="acbbbd49-2b07-4978-8462-dc8ce16d82b0" containerName="registry-server" Dec 05 11:30:01.167508 master-0 kubenswrapper[24928]: I1205 11:30:01.167472 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:01.169679 master-0 kubenswrapper[24928]: I1205 11:30:01.169630 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-fdxl6" Dec 05 11:30:01.171336 master-0 kubenswrapper[24928]: I1205 11:30:01.171301 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 11:30:01.214744 master-0 kubenswrapper[24928]: I1205 11:30:01.214659 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv"] Dec 05 11:30:01.264910 master-0 kubenswrapper[24928]: I1205 11:30:01.264795 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4884e38a-736b-4bc7-b68d-a276eb868adc-config-volume\") pod \"collect-profiles-29415570-f4jrv\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:01.265269 master-0 kubenswrapper[24928]: I1205 11:30:01.264989 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4884e38a-736b-4bc7-b68d-a276eb868adc-secret-volume\") pod \"collect-profiles-29415570-f4jrv\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:01.265269 master-0 kubenswrapper[24928]: I1205 11:30:01.265209 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnsxr\" (UniqueName: \"kubernetes.io/projected/4884e38a-736b-4bc7-b68d-a276eb868adc-kube-api-access-bnsxr\") pod \"collect-profiles-29415570-f4jrv\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:01.369526 master-0 kubenswrapper[24928]: I1205 11:30:01.369448 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4884e38a-736b-4bc7-b68d-a276eb868adc-secret-volume\") pod \"collect-profiles-29415570-f4jrv\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:01.369745 master-0 kubenswrapper[24928]: I1205 11:30:01.369561 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnsxr\" (UniqueName: \"kubernetes.io/projected/4884e38a-736b-4bc7-b68d-a276eb868adc-kube-api-access-bnsxr\") pod \"collect-profiles-29415570-f4jrv\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:01.370005 master-0 kubenswrapper[24928]: I1205 11:30:01.369978 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4884e38a-736b-4bc7-b68d-a276eb868adc-config-volume\") pod \"collect-profiles-29415570-f4jrv\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:01.371681 master-0 kubenswrapper[24928]: I1205 11:30:01.371222 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4884e38a-736b-4bc7-b68d-a276eb868adc-config-volume\") pod \"collect-profiles-29415570-f4jrv\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:01.374822 master-0 kubenswrapper[24928]: I1205 11:30:01.374777 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4884e38a-736b-4bc7-b68d-a276eb868adc-secret-volume\") pod \"collect-profiles-29415570-f4jrv\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:01.386284 master-0 kubenswrapper[24928]: I1205 11:30:01.386190 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnsxr\" (UniqueName: \"kubernetes.io/projected/4884e38a-736b-4bc7-b68d-a276eb868adc-kube-api-access-bnsxr\") pod \"collect-profiles-29415570-f4jrv\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:01.511361 master-0 kubenswrapper[24928]: I1205 11:30:01.511279 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:01.982050 master-0 kubenswrapper[24928]: I1205 11:30:01.981976 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv"] Dec 05 11:30:01.984353 master-0 kubenswrapper[24928]: W1205 11:30:01.984267 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4884e38a_736b_4bc7_b68d_a276eb868adc.slice/crio-c80f6ceaeea8e3576966bfa1ac3b4564527350373fba10c4ac5e194bda4f0f3c WatchSource:0}: Error finding container c80f6ceaeea8e3576966bfa1ac3b4564527350373fba10c4ac5e194bda4f0f3c: Status 404 returned error can't find the container with id c80f6ceaeea8e3576966bfa1ac3b4564527350373fba10c4ac5e194bda4f0f3c Dec 05 11:30:02.551593 master-0 kubenswrapper[24928]: I1205 11:30:02.551383 24928 generic.go:334] "Generic (PLEG): container finished" podID="4884e38a-736b-4bc7-b68d-a276eb868adc" containerID="e8fae3eb69b223f99e393632d8ef1342fb32b3b3b7ca4f4fd77e0275e4197827" exitCode=0 Dec 05 11:30:02.552336 master-0 kubenswrapper[24928]: I1205 11:30:02.551512 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" event={"ID":"4884e38a-736b-4bc7-b68d-a276eb868adc","Type":"ContainerDied","Data":"e8fae3eb69b223f99e393632d8ef1342fb32b3b3b7ca4f4fd77e0275e4197827"} Dec 05 11:30:02.552336 master-0 kubenswrapper[24928]: I1205 11:30:02.551668 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" event={"ID":"4884e38a-736b-4bc7-b68d-a276eb868adc","Type":"ContainerStarted","Data":"c80f6ceaeea8e3576966bfa1ac3b4564527350373fba10c4ac5e194bda4f0f3c"} Dec 05 11:30:04.010712 master-0 kubenswrapper[24928]: I1205 11:30:04.010632 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:04.136214 master-0 kubenswrapper[24928]: I1205 11:30:04.136131 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnsxr\" (UniqueName: \"kubernetes.io/projected/4884e38a-736b-4bc7-b68d-a276eb868adc-kube-api-access-bnsxr\") pod \"4884e38a-736b-4bc7-b68d-a276eb868adc\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " Dec 05 11:30:04.136554 master-0 kubenswrapper[24928]: I1205 11:30:04.136321 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4884e38a-736b-4bc7-b68d-a276eb868adc-secret-volume\") pod \"4884e38a-736b-4bc7-b68d-a276eb868adc\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " Dec 05 11:30:04.136554 master-0 kubenswrapper[24928]: I1205 11:30:04.136391 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4884e38a-736b-4bc7-b68d-a276eb868adc-config-volume\") pod \"4884e38a-736b-4bc7-b68d-a276eb868adc\" (UID: \"4884e38a-736b-4bc7-b68d-a276eb868adc\") " Dec 05 11:30:04.137447 master-0 kubenswrapper[24928]: I1205 11:30:04.137388 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4884e38a-736b-4bc7-b68d-a276eb868adc-config-volume" (OuterVolumeSpecName: "config-volume") pod "4884e38a-736b-4bc7-b68d-a276eb868adc" (UID: "4884e38a-736b-4bc7-b68d-a276eb868adc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:30:04.137658 master-0 kubenswrapper[24928]: I1205 11:30:04.137610 24928 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4884e38a-736b-4bc7-b68d-a276eb868adc-config-volume\") on node \"master-0\" DevicePath \"\"" Dec 05 11:30:04.139815 master-0 kubenswrapper[24928]: I1205 11:30:04.139752 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4884e38a-736b-4bc7-b68d-a276eb868adc-kube-api-access-bnsxr" (OuterVolumeSpecName: "kube-api-access-bnsxr") pod "4884e38a-736b-4bc7-b68d-a276eb868adc" (UID: "4884e38a-736b-4bc7-b68d-a276eb868adc"). InnerVolumeSpecName "kube-api-access-bnsxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:30:04.141064 master-0 kubenswrapper[24928]: I1205 11:30:04.140973 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4884e38a-736b-4bc7-b68d-a276eb868adc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4884e38a-736b-4bc7-b68d-a276eb868adc" (UID: "4884e38a-736b-4bc7-b68d-a276eb868adc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:30:04.239784 master-0 kubenswrapper[24928]: I1205 11:30:04.239704 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnsxr\" (UniqueName: \"kubernetes.io/projected/4884e38a-736b-4bc7-b68d-a276eb868adc-kube-api-access-bnsxr\") on node \"master-0\" DevicePath \"\"" Dec 05 11:30:04.239784 master-0 kubenswrapper[24928]: I1205 11:30:04.239750 24928 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4884e38a-736b-4bc7-b68d-a276eb868adc-secret-volume\") on node \"master-0\" DevicePath \"\"" Dec 05 11:30:04.575783 master-0 kubenswrapper[24928]: I1205 11:30:04.575709 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" event={"ID":"4884e38a-736b-4bc7-b68d-a276eb868adc","Type":"ContainerDied","Data":"c80f6ceaeea8e3576966bfa1ac3b4564527350373fba10c4ac5e194bda4f0f3c"} Dec 05 11:30:04.575783 master-0 kubenswrapper[24928]: I1205 11:30:04.575769 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c80f6ceaeea8e3576966bfa1ac3b4564527350373fba10c4ac5e194bda4f0f3c" Dec 05 11:30:04.576074 master-0 kubenswrapper[24928]: I1205 11:30:04.575881 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415570-f4jrv" Dec 05 11:30:05.190504 master-0 kubenswrapper[24928]: I1205 11:30:05.190466 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7"] Dec 05 11:30:05.214456 master-0 kubenswrapper[24928]: I1205 11:30:05.214282 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415525-82cr7"] Dec 05 11:30:06.223984 master-0 kubenswrapper[24928]: I1205 11:30:06.223924 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7db4c690-9d13-4c82-9565-f974d554916b" path="/var/lib/kubelet/pods/7db4c690-9d13-4c82-9565-f974d554916b/volumes" Dec 05 11:30:49.230331 master-0 kubenswrapper[24928]: I1205 11:30:49.230237 24928 scope.go:117] "RemoveContainer" containerID="2abcdc2c255198aabeec8ea169e4ae237c27c37467559f6c930cf9b54982525d" Dec 05 11:31:24.106809 master-0 kubenswrapper[24928]: I1205 11:31:24.106748 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zk46k"] Dec 05 11:31:24.107390 master-0 kubenswrapper[24928]: E1205 11:31:24.107363 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4884e38a-736b-4bc7-b68d-a276eb868adc" containerName="collect-profiles" Dec 05 11:31:24.107390 master-0 kubenswrapper[24928]: I1205 11:31:24.107384 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="4884e38a-736b-4bc7-b68d-a276eb868adc" containerName="collect-profiles" Dec 05 11:31:24.107860 master-0 kubenswrapper[24928]: I1205 11:31:24.107834 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="4884e38a-736b-4bc7-b68d-a276eb868adc" containerName="collect-profiles" Dec 05 11:31:24.109605 master-0 kubenswrapper[24928]: I1205 11:31:24.109568 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:24.129118 master-0 kubenswrapper[24928]: I1205 11:31:24.129053 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zk46k"] Dec 05 11:31:24.272127 master-0 kubenswrapper[24928]: I1205 11:31:24.271898 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-catalog-content\") pod \"community-operators-zk46k\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:24.276353 master-0 kubenswrapper[24928]: I1205 11:31:24.276266 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-utilities\") pod \"community-operators-zk46k\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:24.276595 master-0 kubenswrapper[24928]: I1205 11:31:24.276567 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thxdb\" (UniqueName: \"kubernetes.io/projected/574befb3-0f24-40c7-b0ba-289e59e4dd69-kube-api-access-thxdb\") pod \"community-operators-zk46k\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:24.379056 master-0 kubenswrapper[24928]: I1205 11:31:24.378971 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thxdb\" (UniqueName: \"kubernetes.io/projected/574befb3-0f24-40c7-b0ba-289e59e4dd69-kube-api-access-thxdb\") pod \"community-operators-zk46k\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:24.382151 master-0 kubenswrapper[24928]: I1205 11:31:24.382072 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-catalog-content\") pod \"community-operators-zk46k\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:24.382351 master-0 kubenswrapper[24928]: I1205 11:31:24.382242 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-utilities\") pod \"community-operators-zk46k\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:24.383337 master-0 kubenswrapper[24928]: I1205 11:31:24.382808 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-catalog-content\") pod \"community-operators-zk46k\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:24.383337 master-0 kubenswrapper[24928]: I1205 11:31:24.382860 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-utilities\") pod \"community-operators-zk46k\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:24.413397 master-0 kubenswrapper[24928]: I1205 11:31:24.413341 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thxdb\" (UniqueName: \"kubernetes.io/projected/574befb3-0f24-40c7-b0ba-289e59e4dd69-kube-api-access-thxdb\") pod \"community-operators-zk46k\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:24.466414 master-0 kubenswrapper[24928]: I1205 11:31:24.466349 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:26.580538 master-0 kubenswrapper[24928]: I1205 11:31:26.580449 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zk46k"] Dec 05 11:31:26.679286 master-0 kubenswrapper[24928]: I1205 11:31:26.679138 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk46k" event={"ID":"574befb3-0f24-40c7-b0ba-289e59e4dd69","Type":"ContainerStarted","Data":"7041064182f182f85667c13b283935a7ba3f6c9db5d723a71739a11b3594acb1"} Dec 05 11:31:27.694133 master-0 kubenswrapper[24928]: I1205 11:31:27.693974 24928 generic.go:334] "Generic (PLEG): container finished" podID="574befb3-0f24-40c7-b0ba-289e59e4dd69" containerID="7261db05f411d8f9276087ff04a310829dae038016129fa6c3ff244dd2138c40" exitCode=0 Dec 05 11:31:27.694133 master-0 kubenswrapper[24928]: I1205 11:31:27.694041 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk46k" event={"ID":"574befb3-0f24-40c7-b0ba-289e59e4dd69","Type":"ContainerDied","Data":"7261db05f411d8f9276087ff04a310829dae038016129fa6c3ff244dd2138c40"} Dec 05 11:31:27.696067 master-0 kubenswrapper[24928]: I1205 11:31:27.695996 24928 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:31:29.724225 master-0 kubenswrapper[24928]: I1205 11:31:29.724141 24928 generic.go:334] "Generic (PLEG): container finished" podID="574befb3-0f24-40c7-b0ba-289e59e4dd69" containerID="4660ed65b769f3ba5cc682e3125c0a19bf16ccadbff434e0362374fd30f50d12" exitCode=0 Dec 05 11:31:29.725228 master-0 kubenswrapper[24928]: I1205 11:31:29.724207 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk46k" event={"ID":"574befb3-0f24-40c7-b0ba-289e59e4dd69","Type":"ContainerDied","Data":"4660ed65b769f3ba5cc682e3125c0a19bf16ccadbff434e0362374fd30f50d12"} Dec 05 11:31:30.742726 master-0 kubenswrapper[24928]: I1205 11:31:30.742660 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk46k" event={"ID":"574befb3-0f24-40c7-b0ba-289e59e4dd69","Type":"ContainerStarted","Data":"e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49"} Dec 05 11:31:30.780170 master-0 kubenswrapper[24928]: I1205 11:31:30.780056 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zk46k" podStartSLOduration=4.231937125 podStartE2EDuration="6.778401445s" podCreationTimestamp="2025-12-05 11:31:24 +0000 UTC" firstStartedPulling="2025-12-05 11:31:27.695790993 +0000 UTC m=+2647.698984884" lastFinishedPulling="2025-12-05 11:31:30.242255353 +0000 UTC m=+2650.245449204" observedRunningTime="2025-12-05 11:31:30.770132229 +0000 UTC m=+2650.773326090" watchObservedRunningTime="2025-12-05 11:31:30.778401445 +0000 UTC m=+2650.781595296" Dec 05 11:31:34.467800 master-0 kubenswrapper[24928]: I1205 11:31:34.467591 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:34.467800 master-0 kubenswrapper[24928]: I1205 11:31:34.467655 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:34.522691 master-0 kubenswrapper[24928]: I1205 11:31:34.522619 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:44.534073 master-0 kubenswrapper[24928]: I1205 11:31:44.533987 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:44.969396 master-0 kubenswrapper[24928]: I1205 11:31:44.969239 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zk46k"] Dec 05 11:31:44.969989 master-0 kubenswrapper[24928]: I1205 11:31:44.969928 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zk46k" podUID="574befb3-0f24-40c7-b0ba-289e59e4dd69" containerName="registry-server" containerID="cri-o://e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49" gracePeriod=2 Dec 05 11:31:45.469459 master-0 kubenswrapper[24928]: I1205 11:31:45.469393 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:45.583468 master-0 kubenswrapper[24928]: I1205 11:31:45.583389 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-catalog-content\") pod \"574befb3-0f24-40c7-b0ba-289e59e4dd69\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " Dec 05 11:31:45.584247 master-0 kubenswrapper[24928]: I1205 11:31:45.583766 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thxdb\" (UniqueName: \"kubernetes.io/projected/574befb3-0f24-40c7-b0ba-289e59e4dd69-kube-api-access-thxdb\") pod \"574befb3-0f24-40c7-b0ba-289e59e4dd69\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " Dec 05 11:31:45.584247 master-0 kubenswrapper[24928]: I1205 11:31:45.583887 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-utilities\") pod \"574befb3-0f24-40c7-b0ba-289e59e4dd69\" (UID: \"574befb3-0f24-40c7-b0ba-289e59e4dd69\") " Dec 05 11:31:45.586190 master-0 kubenswrapper[24928]: I1205 11:31:45.586144 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-utilities" (OuterVolumeSpecName: "utilities") pod "574befb3-0f24-40c7-b0ba-289e59e4dd69" (UID: "574befb3-0f24-40c7-b0ba-289e59e4dd69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:45.596459 master-0 kubenswrapper[24928]: I1205 11:31:45.596322 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/574befb3-0f24-40c7-b0ba-289e59e4dd69-kube-api-access-thxdb" (OuterVolumeSpecName: "kube-api-access-thxdb") pod "574befb3-0f24-40c7-b0ba-289e59e4dd69" (UID: "574befb3-0f24-40c7-b0ba-289e59e4dd69"). InnerVolumeSpecName "kube-api-access-thxdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:31:45.656679 master-0 kubenswrapper[24928]: I1205 11:31:45.656628 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "574befb3-0f24-40c7-b0ba-289e59e4dd69" (UID: "574befb3-0f24-40c7-b0ba-289e59e4dd69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:31:45.690632 master-0 kubenswrapper[24928]: I1205 11:31:45.690544 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:31:45.690632 master-0 kubenswrapper[24928]: I1205 11:31:45.690613 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thxdb\" (UniqueName: \"kubernetes.io/projected/574befb3-0f24-40c7-b0ba-289e59e4dd69-kube-api-access-thxdb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:31:45.690632 master-0 kubenswrapper[24928]: I1205 11:31:45.690631 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574befb3-0f24-40c7-b0ba-289e59e4dd69-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:31:45.976234 master-0 kubenswrapper[24928]: I1205 11:31:45.976052 24928 generic.go:334] "Generic (PLEG): container finished" podID="574befb3-0f24-40c7-b0ba-289e59e4dd69" containerID="e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49" exitCode=0 Dec 05 11:31:45.976234 master-0 kubenswrapper[24928]: I1205 11:31:45.976123 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zk46k" Dec 05 11:31:45.976695 master-0 kubenswrapper[24928]: I1205 11:31:45.976140 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk46k" event={"ID":"574befb3-0f24-40c7-b0ba-289e59e4dd69","Type":"ContainerDied","Data":"e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49"} Dec 05 11:31:45.976695 master-0 kubenswrapper[24928]: I1205 11:31:45.976296 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zk46k" event={"ID":"574befb3-0f24-40c7-b0ba-289e59e4dd69","Type":"ContainerDied","Data":"7041064182f182f85667c13b283935a7ba3f6c9db5d723a71739a11b3594acb1"} Dec 05 11:31:45.976695 master-0 kubenswrapper[24928]: I1205 11:31:45.976320 24928 scope.go:117] "RemoveContainer" containerID="e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49" Dec 05 11:31:46.010659 master-0 kubenswrapper[24928]: I1205 11:31:46.010579 24928 scope.go:117] "RemoveContainer" containerID="4660ed65b769f3ba5cc682e3125c0a19bf16ccadbff434e0362374fd30f50d12" Dec 05 11:31:46.048221 master-0 kubenswrapper[24928]: I1205 11:31:46.048114 24928 scope.go:117] "RemoveContainer" containerID="7261db05f411d8f9276087ff04a310829dae038016129fa6c3ff244dd2138c40" Dec 05 11:31:46.103296 master-0 kubenswrapper[24928]: I1205 11:31:46.103238 24928 scope.go:117] "RemoveContainer" containerID="e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49" Dec 05 11:31:46.103830 master-0 kubenswrapper[24928]: E1205 11:31:46.103751 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49\": container with ID starting with e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49 not found: ID does not exist" containerID="e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49" Dec 05 11:31:46.103915 master-0 kubenswrapper[24928]: I1205 11:31:46.103828 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49"} err="failed to get container status \"e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49\": rpc error: code = NotFound desc = could not find container \"e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49\": container with ID starting with e723be159343496b55bd39366f4da3a6215103d6765c6acca3768576465c6b49 not found: ID does not exist" Dec 05 11:31:46.103915 master-0 kubenswrapper[24928]: I1205 11:31:46.103871 24928 scope.go:117] "RemoveContainer" containerID="4660ed65b769f3ba5cc682e3125c0a19bf16ccadbff434e0362374fd30f50d12" Dec 05 11:31:46.104367 master-0 kubenswrapper[24928]: E1205 11:31:46.104324 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4660ed65b769f3ba5cc682e3125c0a19bf16ccadbff434e0362374fd30f50d12\": container with ID starting with 4660ed65b769f3ba5cc682e3125c0a19bf16ccadbff434e0362374fd30f50d12 not found: ID does not exist" containerID="4660ed65b769f3ba5cc682e3125c0a19bf16ccadbff434e0362374fd30f50d12" Dec 05 11:31:46.104586 master-0 kubenswrapper[24928]: I1205 11:31:46.104367 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4660ed65b769f3ba5cc682e3125c0a19bf16ccadbff434e0362374fd30f50d12"} err="failed to get container status \"4660ed65b769f3ba5cc682e3125c0a19bf16ccadbff434e0362374fd30f50d12\": rpc error: code = NotFound desc = could not find container \"4660ed65b769f3ba5cc682e3125c0a19bf16ccadbff434e0362374fd30f50d12\": container with ID starting with 4660ed65b769f3ba5cc682e3125c0a19bf16ccadbff434e0362374fd30f50d12 not found: ID does not exist" Dec 05 11:31:46.104586 master-0 kubenswrapper[24928]: I1205 11:31:46.104397 24928 scope.go:117] "RemoveContainer" containerID="7261db05f411d8f9276087ff04a310829dae038016129fa6c3ff244dd2138c40" Dec 05 11:31:46.104903 master-0 kubenswrapper[24928]: E1205 11:31:46.104860 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7261db05f411d8f9276087ff04a310829dae038016129fa6c3ff244dd2138c40\": container with ID starting with 7261db05f411d8f9276087ff04a310829dae038016129fa6c3ff244dd2138c40 not found: ID does not exist" containerID="7261db05f411d8f9276087ff04a310829dae038016129fa6c3ff244dd2138c40" Dec 05 11:31:46.104903 master-0 kubenswrapper[24928]: I1205 11:31:46.104889 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7261db05f411d8f9276087ff04a310829dae038016129fa6c3ff244dd2138c40"} err="failed to get container status \"7261db05f411d8f9276087ff04a310829dae038016129fa6c3ff244dd2138c40\": rpc error: code = NotFound desc = could not find container \"7261db05f411d8f9276087ff04a310829dae038016129fa6c3ff244dd2138c40\": container with ID starting with 7261db05f411d8f9276087ff04a310829dae038016129fa6c3ff244dd2138c40 not found: ID does not exist" Dec 05 11:31:46.310217 master-0 kubenswrapper[24928]: I1205 11:31:46.310140 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zk46k"] Dec 05 11:31:46.322832 master-0 kubenswrapper[24928]: I1205 11:31:46.322756 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zk46k"] Dec 05 11:31:48.222845 master-0 kubenswrapper[24928]: I1205 11:31:48.222716 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="574befb3-0f24-40c7-b0ba-289e59e4dd69" path="/var/lib/kubelet/pods/574befb3-0f24-40c7-b0ba-289e59e4dd69/volumes" Dec 05 11:33:21.514040 master-0 kubenswrapper[24928]: I1205 11:33:21.513970 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mpzmp"] Dec 05 11:33:21.514786 master-0 kubenswrapper[24928]: E1205 11:33:21.514752 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574befb3-0f24-40c7-b0ba-289e59e4dd69" containerName="extract-content" Dec 05 11:33:21.514786 master-0 kubenswrapper[24928]: I1205 11:33:21.514773 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="574befb3-0f24-40c7-b0ba-289e59e4dd69" containerName="extract-content" Dec 05 11:33:21.514892 master-0 kubenswrapper[24928]: E1205 11:33:21.514797 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574befb3-0f24-40c7-b0ba-289e59e4dd69" containerName="extract-utilities" Dec 05 11:33:21.514892 master-0 kubenswrapper[24928]: I1205 11:33:21.514806 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="574befb3-0f24-40c7-b0ba-289e59e4dd69" containerName="extract-utilities" Dec 05 11:33:21.514892 master-0 kubenswrapper[24928]: E1205 11:33:21.514846 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574befb3-0f24-40c7-b0ba-289e59e4dd69" containerName="registry-server" Dec 05 11:33:21.514892 master-0 kubenswrapper[24928]: I1205 11:33:21.514855 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="574befb3-0f24-40c7-b0ba-289e59e4dd69" containerName="registry-server" Dec 05 11:33:21.515230 master-0 kubenswrapper[24928]: I1205 11:33:21.515202 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="574befb3-0f24-40c7-b0ba-289e59e4dd69" containerName="registry-server" Dec 05 11:33:21.516988 master-0 kubenswrapper[24928]: I1205 11:33:21.516958 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:21.564818 master-0 kubenswrapper[24928]: I1205 11:33:21.562248 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpzmp"] Dec 05 11:33:21.663869 master-0 kubenswrapper[24928]: I1205 11:33:21.663793 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/535b1b4c-2431-4632-b095-9b46bcf6afb3-utilities\") pod \"redhat-marketplace-mpzmp\" (UID: \"535b1b4c-2431-4632-b095-9b46bcf6afb3\") " pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:21.665237 master-0 kubenswrapper[24928]: I1205 11:33:21.665124 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/535b1b4c-2431-4632-b095-9b46bcf6afb3-catalog-content\") pod \"redhat-marketplace-mpzmp\" (UID: \"535b1b4c-2431-4632-b095-9b46bcf6afb3\") " pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:21.665342 master-0 kubenswrapper[24928]: I1205 11:33:21.665255 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdjgw\" (UniqueName: \"kubernetes.io/projected/535b1b4c-2431-4632-b095-9b46bcf6afb3-kube-api-access-kdjgw\") pod \"redhat-marketplace-mpzmp\" (UID: \"535b1b4c-2431-4632-b095-9b46bcf6afb3\") " pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:21.701111 master-0 kubenswrapper[24928]: I1205 11:33:21.701022 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7xtcd"] Dec 05 11:33:21.704229 master-0 kubenswrapper[24928]: I1205 11:33:21.704172 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:21.712031 master-0 kubenswrapper[24928]: I1205 11:33:21.711950 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7xtcd"] Dec 05 11:33:21.767545 master-0 kubenswrapper[24928]: I1205 11:33:21.767252 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/535b1b4c-2431-4632-b095-9b46bcf6afb3-catalog-content\") pod \"redhat-marketplace-mpzmp\" (UID: \"535b1b4c-2431-4632-b095-9b46bcf6afb3\") " pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:21.767545 master-0 kubenswrapper[24928]: I1205 11:33:21.767328 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdjgw\" (UniqueName: \"kubernetes.io/projected/535b1b4c-2431-4632-b095-9b46bcf6afb3-kube-api-access-kdjgw\") pod \"redhat-marketplace-mpzmp\" (UID: \"535b1b4c-2431-4632-b095-9b46bcf6afb3\") " pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:21.767545 master-0 kubenswrapper[24928]: I1205 11:33:21.767365 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/535b1b4c-2431-4632-b095-9b46bcf6afb3-utilities\") pod \"redhat-marketplace-mpzmp\" (UID: \"535b1b4c-2431-4632-b095-9b46bcf6afb3\") " pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:21.768270 master-0 kubenswrapper[24928]: I1205 11:33:21.768076 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/535b1b4c-2431-4632-b095-9b46bcf6afb3-utilities\") pod \"redhat-marketplace-mpzmp\" (UID: \"535b1b4c-2431-4632-b095-9b46bcf6afb3\") " pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:21.775800 master-0 kubenswrapper[24928]: I1205 11:33:21.768323 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/535b1b4c-2431-4632-b095-9b46bcf6afb3-catalog-content\") pod \"redhat-marketplace-mpzmp\" (UID: \"535b1b4c-2431-4632-b095-9b46bcf6afb3\") " pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:21.794356 master-0 kubenswrapper[24928]: I1205 11:33:21.793895 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdjgw\" (UniqueName: \"kubernetes.io/projected/535b1b4c-2431-4632-b095-9b46bcf6afb3-kube-api-access-kdjgw\") pod \"redhat-marketplace-mpzmp\" (UID: \"535b1b4c-2431-4632-b095-9b46bcf6afb3\") " pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:21.850577 master-0 kubenswrapper[24928]: I1205 11:33:21.848880 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:21.872449 master-0 kubenswrapper[24928]: I1205 11:33:21.869674 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-utilities\") pod \"redhat-operators-7xtcd\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:21.872449 master-0 kubenswrapper[24928]: I1205 11:33:21.869770 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95pxq\" (UniqueName: \"kubernetes.io/projected/8a41b09b-681e-423e-8b02-aa19271b0135-kube-api-access-95pxq\") pod \"redhat-operators-7xtcd\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:21.872449 master-0 kubenswrapper[24928]: I1205 11:33:21.869876 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-catalog-content\") pod \"redhat-operators-7xtcd\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:21.973262 master-0 kubenswrapper[24928]: I1205 11:33:21.973185 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-utilities\") pod \"redhat-operators-7xtcd\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:21.973503 master-0 kubenswrapper[24928]: I1205 11:33:21.973289 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95pxq\" (UniqueName: \"kubernetes.io/projected/8a41b09b-681e-423e-8b02-aa19271b0135-kube-api-access-95pxq\") pod \"redhat-operators-7xtcd\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:21.973503 master-0 kubenswrapper[24928]: I1205 11:33:21.973391 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-catalog-content\") pod \"redhat-operators-7xtcd\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:21.977440 master-0 kubenswrapper[24928]: I1205 11:33:21.974094 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-catalog-content\") pod \"redhat-operators-7xtcd\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:21.977440 master-0 kubenswrapper[24928]: I1205 11:33:21.974587 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-utilities\") pod \"redhat-operators-7xtcd\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:21.999398 master-0 kubenswrapper[24928]: I1205 11:33:21.995492 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95pxq\" (UniqueName: \"kubernetes.io/projected/8a41b09b-681e-423e-8b02-aa19271b0135-kube-api-access-95pxq\") pod \"redhat-operators-7xtcd\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:22.046651 master-0 kubenswrapper[24928]: I1205 11:33:22.041971 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:22.419534 master-0 kubenswrapper[24928]: I1205 11:33:22.414721 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpzmp"] Dec 05 11:33:22.592687 master-0 kubenswrapper[24928]: I1205 11:33:22.592580 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7xtcd"] Dec 05 11:33:23.367308 master-0 kubenswrapper[24928]: I1205 11:33:23.367177 24928 generic.go:334] "Generic (PLEG): container finished" podID="8a41b09b-681e-423e-8b02-aa19271b0135" containerID="b5eca4d59e53d2bca5071770d226588a253426de3d78bd3f01179900fde33806" exitCode=0 Dec 05 11:33:23.367308 master-0 kubenswrapper[24928]: I1205 11:33:23.367274 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xtcd" event={"ID":"8a41b09b-681e-423e-8b02-aa19271b0135","Type":"ContainerDied","Data":"b5eca4d59e53d2bca5071770d226588a253426de3d78bd3f01179900fde33806"} Dec 05 11:33:23.367308 master-0 kubenswrapper[24928]: I1205 11:33:23.367309 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xtcd" event={"ID":"8a41b09b-681e-423e-8b02-aa19271b0135","Type":"ContainerStarted","Data":"eb73419351be705d97ca593f28e2345b8357c0ce8289fe7d4c0c7691e113dc03"} Dec 05 11:33:23.372474 master-0 kubenswrapper[24928]: I1205 11:33:23.371250 24928 generic.go:334] "Generic (PLEG): container finished" podID="535b1b4c-2431-4632-b095-9b46bcf6afb3" containerID="fda7c2d187b635dee9a72f93d0ba1bff2c681ddf9eaf4ca5e83d12f2e09720fd" exitCode=0 Dec 05 11:33:23.372474 master-0 kubenswrapper[24928]: I1205 11:33:23.371856 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzmp" event={"ID":"535b1b4c-2431-4632-b095-9b46bcf6afb3","Type":"ContainerDied","Data":"fda7c2d187b635dee9a72f93d0ba1bff2c681ddf9eaf4ca5e83d12f2e09720fd"} Dec 05 11:33:23.372716 master-0 kubenswrapper[24928]: I1205 11:33:23.372496 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzmp" event={"ID":"535b1b4c-2431-4632-b095-9b46bcf6afb3","Type":"ContainerStarted","Data":"4fcfbde5f16fef8198a400fe483a40b80a499dd3b77c69124b19f27a2fe494cb"} Dec 05 11:33:24.393916 master-0 kubenswrapper[24928]: I1205 11:33:24.393854 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xtcd" event={"ID":"8a41b09b-681e-423e-8b02-aa19271b0135","Type":"ContainerStarted","Data":"872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24"} Dec 05 11:33:25.416854 master-0 kubenswrapper[24928]: I1205 11:33:25.416782 24928 generic.go:334] "Generic (PLEG): container finished" podID="8a41b09b-681e-423e-8b02-aa19271b0135" containerID="872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24" exitCode=0 Dec 05 11:33:25.416854 master-0 kubenswrapper[24928]: I1205 11:33:25.416848 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xtcd" event={"ID":"8a41b09b-681e-423e-8b02-aa19271b0135","Type":"ContainerDied","Data":"872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24"} Dec 05 11:33:26.446406 master-0 kubenswrapper[24928]: I1205 11:33:26.446340 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzmp" event={"ID":"535b1b4c-2431-4632-b095-9b46bcf6afb3","Type":"ContainerStarted","Data":"9165aa1ed893a0fa13562b584f14911fcf11d30b565d953d1cc228e6ab47b7fe"} Dec 05 11:33:27.459593 master-0 kubenswrapper[24928]: I1205 11:33:27.459518 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xtcd" event={"ID":"8a41b09b-681e-423e-8b02-aa19271b0135","Type":"ContainerStarted","Data":"e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22"} Dec 05 11:33:27.463017 master-0 kubenswrapper[24928]: I1205 11:33:27.461930 24928 generic.go:334] "Generic (PLEG): container finished" podID="535b1b4c-2431-4632-b095-9b46bcf6afb3" containerID="9165aa1ed893a0fa13562b584f14911fcf11d30b565d953d1cc228e6ab47b7fe" exitCode=0 Dec 05 11:33:27.463017 master-0 kubenswrapper[24928]: I1205 11:33:27.461990 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzmp" event={"ID":"535b1b4c-2431-4632-b095-9b46bcf6afb3","Type":"ContainerDied","Data":"9165aa1ed893a0fa13562b584f14911fcf11d30b565d953d1cc228e6ab47b7fe"} Dec 05 11:33:27.496273 master-0 kubenswrapper[24928]: I1205 11:33:27.496181 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7xtcd" podStartSLOduration=3.344688708 podStartE2EDuration="6.496156884s" podCreationTimestamp="2025-12-05 11:33:21 +0000 UTC" firstStartedPulling="2025-12-05 11:33:23.369873604 +0000 UTC m=+2763.373067455" lastFinishedPulling="2025-12-05 11:33:26.52134179 +0000 UTC m=+2766.524535631" observedRunningTime="2025-12-05 11:33:27.48139218 +0000 UTC m=+2767.484586061" watchObservedRunningTime="2025-12-05 11:33:27.496156884 +0000 UTC m=+2767.499350735" Dec 05 11:33:28.477143 master-0 kubenswrapper[24928]: I1205 11:33:28.476989 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpzmp" event={"ID":"535b1b4c-2431-4632-b095-9b46bcf6afb3","Type":"ContainerStarted","Data":"4f3dfbe174c1383eb52aa3753a589cca8774c09742e29a3b3d965c89650408fc"} Dec 05 11:33:28.509596 master-0 kubenswrapper[24928]: I1205 11:33:28.509487 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mpzmp" podStartSLOduration=3.003134171 podStartE2EDuration="7.509466266s" podCreationTimestamp="2025-12-05 11:33:21 +0000 UTC" firstStartedPulling="2025-12-05 11:33:23.372990371 +0000 UTC m=+2763.376184222" lastFinishedPulling="2025-12-05 11:33:27.879322466 +0000 UTC m=+2767.882516317" observedRunningTime="2025-12-05 11:33:28.499458219 +0000 UTC m=+2768.502652090" watchObservedRunningTime="2025-12-05 11:33:28.509466266 +0000 UTC m=+2768.512660117" Dec 05 11:33:31.849932 master-0 kubenswrapper[24928]: I1205 11:33:31.849847 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:31.849932 master-0 kubenswrapper[24928]: I1205 11:33:31.849919 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:31.905096 master-0 kubenswrapper[24928]: I1205 11:33:31.905028 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:32.043822 master-0 kubenswrapper[24928]: I1205 11:33:32.043735 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:32.043822 master-0 kubenswrapper[24928]: I1205 11:33:32.043814 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:32.092980 master-0 kubenswrapper[24928]: I1205 11:33:32.092904 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:32.582291 master-0 kubenswrapper[24928]: I1205 11:33:32.582217 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:34.082499 master-0 kubenswrapper[24928]: I1205 11:33:34.082397 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7xtcd"] Dec 05 11:33:34.561516 master-0 kubenswrapper[24928]: I1205 11:33:34.561457 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7xtcd" podUID="8a41b09b-681e-423e-8b02-aa19271b0135" containerName="registry-server" containerID="cri-o://e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22" gracePeriod=2 Dec 05 11:33:35.128328 master-0 kubenswrapper[24928]: I1205 11:33:35.128263 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:35.270413 master-0 kubenswrapper[24928]: I1205 11:33:35.270308 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-utilities\") pod \"8a41b09b-681e-423e-8b02-aa19271b0135\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " Dec 05 11:33:35.271053 master-0 kubenswrapper[24928]: I1205 11:33:35.271019 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95pxq\" (UniqueName: \"kubernetes.io/projected/8a41b09b-681e-423e-8b02-aa19271b0135-kube-api-access-95pxq\") pod \"8a41b09b-681e-423e-8b02-aa19271b0135\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " Dec 05 11:33:35.271170 master-0 kubenswrapper[24928]: I1205 11:33:35.271154 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-catalog-content\") pod \"8a41b09b-681e-423e-8b02-aa19271b0135\" (UID: \"8a41b09b-681e-423e-8b02-aa19271b0135\") " Dec 05 11:33:35.271273 master-0 kubenswrapper[24928]: I1205 11:33:35.271218 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-utilities" (OuterVolumeSpecName: "utilities") pod "8a41b09b-681e-423e-8b02-aa19271b0135" (UID: "8a41b09b-681e-423e-8b02-aa19271b0135"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:33:35.271884 master-0 kubenswrapper[24928]: I1205 11:33:35.271861 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:33:35.275230 master-0 kubenswrapper[24928]: I1205 11:33:35.275142 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a41b09b-681e-423e-8b02-aa19271b0135-kube-api-access-95pxq" (OuterVolumeSpecName: "kube-api-access-95pxq") pod "8a41b09b-681e-423e-8b02-aa19271b0135" (UID: "8a41b09b-681e-423e-8b02-aa19271b0135"). InnerVolumeSpecName "kube-api-access-95pxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:33:35.373822 master-0 kubenswrapper[24928]: I1205 11:33:35.373765 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95pxq\" (UniqueName: \"kubernetes.io/projected/8a41b09b-681e-423e-8b02-aa19271b0135-kube-api-access-95pxq\") on node \"master-0\" DevicePath \"\"" Dec 05 11:33:35.574721 master-0 kubenswrapper[24928]: I1205 11:33:35.574654 24928 generic.go:334] "Generic (PLEG): container finished" podID="8a41b09b-681e-423e-8b02-aa19271b0135" containerID="e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22" exitCode=0 Dec 05 11:33:35.574721 master-0 kubenswrapper[24928]: I1205 11:33:35.574711 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xtcd" event={"ID":"8a41b09b-681e-423e-8b02-aa19271b0135","Type":"ContainerDied","Data":"e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22"} Dec 05 11:33:35.575033 master-0 kubenswrapper[24928]: I1205 11:33:35.574726 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7xtcd" Dec 05 11:33:35.575033 master-0 kubenswrapper[24928]: I1205 11:33:35.574751 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7xtcd" event={"ID":"8a41b09b-681e-423e-8b02-aa19271b0135","Type":"ContainerDied","Data":"eb73419351be705d97ca593f28e2345b8357c0ce8289fe7d4c0c7691e113dc03"} Dec 05 11:33:35.575033 master-0 kubenswrapper[24928]: I1205 11:33:35.574776 24928 scope.go:117] "RemoveContainer" containerID="e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22" Dec 05 11:33:35.596569 master-0 kubenswrapper[24928]: I1205 11:33:35.596505 24928 scope.go:117] "RemoveContainer" containerID="872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24" Dec 05 11:33:35.618163 master-0 kubenswrapper[24928]: I1205 11:33:35.618105 24928 scope.go:117] "RemoveContainer" containerID="b5eca4d59e53d2bca5071770d226588a253426de3d78bd3f01179900fde33806" Dec 05 11:33:35.631598 master-0 kubenswrapper[24928]: I1205 11:33:35.631406 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a41b09b-681e-423e-8b02-aa19271b0135" (UID: "8a41b09b-681e-423e-8b02-aa19271b0135"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:33:35.668935 master-0 kubenswrapper[24928]: I1205 11:33:35.668883 24928 scope.go:117] "RemoveContainer" containerID="e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22" Dec 05 11:33:35.669706 master-0 kubenswrapper[24928]: E1205 11:33:35.669645 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22\": container with ID starting with e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22 not found: ID does not exist" containerID="e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22" Dec 05 11:33:35.669795 master-0 kubenswrapper[24928]: I1205 11:33:35.669711 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22"} err="failed to get container status \"e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22\": rpc error: code = NotFound desc = could not find container \"e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22\": container with ID starting with e9b4e7e3910e40b26e5d91331cbe87e54b78544aef2ead267fc71a819ed42c22 not found: ID does not exist" Dec 05 11:33:35.669795 master-0 kubenswrapper[24928]: I1205 11:33:35.669740 24928 scope.go:117] "RemoveContainer" containerID="872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24" Dec 05 11:33:35.670275 master-0 kubenswrapper[24928]: E1205 11:33:35.670242 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24\": container with ID starting with 872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24 not found: ID does not exist" containerID="872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24" Dec 05 11:33:35.670515 master-0 kubenswrapper[24928]: I1205 11:33:35.670478 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24"} err="failed to get container status \"872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24\": rpc error: code = NotFound desc = could not find container \"872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24\": container with ID starting with 872f7a322941a2b299494107834e3ea5365427dc946838b5f913caf178b55d24 not found: ID does not exist" Dec 05 11:33:35.670610 master-0 kubenswrapper[24928]: I1205 11:33:35.670592 24928 scope.go:117] "RemoveContainer" containerID="b5eca4d59e53d2bca5071770d226588a253426de3d78bd3f01179900fde33806" Dec 05 11:33:35.671028 master-0 kubenswrapper[24928]: E1205 11:33:35.670995 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5eca4d59e53d2bca5071770d226588a253426de3d78bd3f01179900fde33806\": container with ID starting with b5eca4d59e53d2bca5071770d226588a253426de3d78bd3f01179900fde33806 not found: ID does not exist" containerID="b5eca4d59e53d2bca5071770d226588a253426de3d78bd3f01179900fde33806" Dec 05 11:33:35.671028 master-0 kubenswrapper[24928]: I1205 11:33:35.671028 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5eca4d59e53d2bca5071770d226588a253426de3d78bd3f01179900fde33806"} err="failed to get container status \"b5eca4d59e53d2bca5071770d226588a253426de3d78bd3f01179900fde33806\": rpc error: code = NotFound desc = could not find container \"b5eca4d59e53d2bca5071770d226588a253426de3d78bd3f01179900fde33806\": container with ID starting with b5eca4d59e53d2bca5071770d226588a253426de3d78bd3f01179900fde33806 not found: ID does not exist" Dec 05 11:33:35.681902 master-0 kubenswrapper[24928]: I1205 11:33:35.681855 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a41b09b-681e-423e-8b02-aa19271b0135-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:33:35.921902 master-0 kubenswrapper[24928]: I1205 11:33:35.921822 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7xtcd"] Dec 05 11:33:35.932350 master-0 kubenswrapper[24928]: I1205 11:33:35.932274 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7xtcd"] Dec 05 11:33:36.218912 master-0 kubenswrapper[24928]: I1205 11:33:36.218810 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a41b09b-681e-423e-8b02-aa19271b0135" path="/var/lib/kubelet/pods/8a41b09b-681e-423e-8b02-aa19271b0135/volumes" Dec 05 11:33:41.901098 master-0 kubenswrapper[24928]: I1205 11:33:41.901033 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mpzmp" Dec 05 11:33:41.988781 master-0 kubenswrapper[24928]: I1205 11:33:41.988687 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpzmp"] Dec 05 11:33:42.035565 master-0 kubenswrapper[24928]: I1205 11:33:42.033584 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wk29h"] Dec 05 11:33:42.035565 master-0 kubenswrapper[24928]: I1205 11:33:42.033905 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wk29h" podUID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerName="registry-server" containerID="cri-o://2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22" gracePeriod=2 Dec 05 11:33:42.596562 master-0 kubenswrapper[24928]: I1205 11:33:42.596498 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 11:33:42.650636 master-0 kubenswrapper[24928]: I1205 11:33:42.650555 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnzgs\" (UniqueName: \"kubernetes.io/projected/0aa6e138-4b1f-4047-8255-a2b14d044588-kube-api-access-wnzgs\") pod \"0aa6e138-4b1f-4047-8255-a2b14d044588\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " Dec 05 11:33:42.650885 master-0 kubenswrapper[24928]: I1205 11:33:42.650739 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-catalog-content\") pod \"0aa6e138-4b1f-4047-8255-a2b14d044588\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " Dec 05 11:33:42.650885 master-0 kubenswrapper[24928]: I1205 11:33:42.650866 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-utilities\") pod \"0aa6e138-4b1f-4047-8255-a2b14d044588\" (UID: \"0aa6e138-4b1f-4047-8255-a2b14d044588\") " Dec 05 11:33:42.656624 master-0 kubenswrapper[24928]: I1205 11:33:42.652151 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-utilities" (OuterVolumeSpecName: "utilities") pod "0aa6e138-4b1f-4047-8255-a2b14d044588" (UID: "0aa6e138-4b1f-4047-8255-a2b14d044588"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:33:42.656624 master-0 kubenswrapper[24928]: I1205 11:33:42.655000 24928 generic.go:334] "Generic (PLEG): container finished" podID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerID="2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22" exitCode=0 Dec 05 11:33:42.656624 master-0 kubenswrapper[24928]: I1205 11:33:42.655116 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wk29h" Dec 05 11:33:42.656624 master-0 kubenswrapper[24928]: I1205 11:33:42.655354 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk29h" event={"ID":"0aa6e138-4b1f-4047-8255-a2b14d044588","Type":"ContainerDied","Data":"2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22"} Dec 05 11:33:42.656624 master-0 kubenswrapper[24928]: I1205 11:33:42.655396 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wk29h" event={"ID":"0aa6e138-4b1f-4047-8255-a2b14d044588","Type":"ContainerDied","Data":"52b24a28414c04f2574070f038561f387ec44118f13f6cac5eab1761c0750f3a"} Dec 05 11:33:42.656624 master-0 kubenswrapper[24928]: I1205 11:33:42.655430 24928 scope.go:117] "RemoveContainer" containerID="2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22" Dec 05 11:33:42.665254 master-0 kubenswrapper[24928]: I1205 11:33:42.662842 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0aa6e138-4b1f-4047-8255-a2b14d044588-kube-api-access-wnzgs" (OuterVolumeSpecName: "kube-api-access-wnzgs") pod "0aa6e138-4b1f-4047-8255-a2b14d044588" (UID: "0aa6e138-4b1f-4047-8255-a2b14d044588"). InnerVolumeSpecName "kube-api-access-wnzgs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:33:42.666385 master-0 kubenswrapper[24928]: I1205 11:33:42.666334 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0aa6e138-4b1f-4047-8255-a2b14d044588" (UID: "0aa6e138-4b1f-4047-8255-a2b14d044588"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:33:42.754773 master-0 kubenswrapper[24928]: I1205 11:33:42.754152 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnzgs\" (UniqueName: \"kubernetes.io/projected/0aa6e138-4b1f-4047-8255-a2b14d044588-kube-api-access-wnzgs\") on node \"master-0\" DevicePath \"\"" Dec 05 11:33:42.754773 master-0 kubenswrapper[24928]: I1205 11:33:42.754201 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:33:42.754773 master-0 kubenswrapper[24928]: I1205 11:33:42.754212 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0aa6e138-4b1f-4047-8255-a2b14d044588-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:33:42.759697 master-0 kubenswrapper[24928]: I1205 11:33:42.759090 24928 scope.go:117] "RemoveContainer" containerID="e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23" Dec 05 11:33:42.806685 master-0 kubenswrapper[24928]: I1205 11:33:42.806632 24928 scope.go:117] "RemoveContainer" containerID="698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90" Dec 05 11:33:42.839945 master-0 kubenswrapper[24928]: I1205 11:33:42.838080 24928 scope.go:117] "RemoveContainer" containerID="2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22" Dec 05 11:33:42.839945 master-0 kubenswrapper[24928]: E1205 11:33:42.838563 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22\": container with ID starting with 2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22 not found: ID does not exist" containerID="2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22" Dec 05 11:33:42.839945 master-0 kubenswrapper[24928]: I1205 11:33:42.838619 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22"} err="failed to get container status \"2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22\": rpc error: code = NotFound desc = could not find container \"2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22\": container with ID starting with 2075bc6940a583b0e399b5d836255da8bf145674cb48718a88fdde4b865bbd22 not found: ID does not exist" Dec 05 11:33:42.839945 master-0 kubenswrapper[24928]: I1205 11:33:42.838654 24928 scope.go:117] "RemoveContainer" containerID="e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23" Dec 05 11:33:42.839945 master-0 kubenswrapper[24928]: E1205 11:33:42.838973 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23\": container with ID starting with e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23 not found: ID does not exist" containerID="e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23" Dec 05 11:33:42.839945 master-0 kubenswrapper[24928]: I1205 11:33:42.839008 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23"} err="failed to get container status \"e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23\": rpc error: code = NotFound desc = could not find container \"e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23\": container with ID starting with e8997fed8ea5d76f9f07664182c6762917274450dcbbcc0d3aa66764c6d81f23 not found: ID does not exist" Dec 05 11:33:42.839945 master-0 kubenswrapper[24928]: I1205 11:33:42.839034 24928 scope.go:117] "RemoveContainer" containerID="698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90" Dec 05 11:33:42.839945 master-0 kubenswrapper[24928]: E1205 11:33:42.839704 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90\": container with ID starting with 698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90 not found: ID does not exist" containerID="698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90" Dec 05 11:33:42.839945 master-0 kubenswrapper[24928]: I1205 11:33:42.839746 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90"} err="failed to get container status \"698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90\": rpc error: code = NotFound desc = could not find container \"698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90\": container with ID starting with 698c728d648f0d4b494c997cec6ab59e4cbdce2d3452fd917b9ac3eb2b778d90 not found: ID does not exist" Dec 05 11:33:43.168044 master-0 kubenswrapper[24928]: I1205 11:33:43.167975 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wk29h"] Dec 05 11:33:43.184690 master-0 kubenswrapper[24928]: I1205 11:33:43.184612 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wk29h"] Dec 05 11:33:44.223007 master-0 kubenswrapper[24928]: I1205 11:33:44.222934 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0aa6e138-4b1f-4047-8255-a2b14d044588" path="/var/lib/kubelet/pods/0aa6e138-4b1f-4047-8255-a2b14d044588/volumes" Dec 05 11:35:58.713177 master-0 kubenswrapper[24928]: I1205 11:35:58.713062 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4mvw4"] Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: E1205 11:35:58.713689 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a41b09b-681e-423e-8b02-aa19271b0135" containerName="extract-content" Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: I1205 11:35:58.713705 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a41b09b-681e-423e-8b02-aa19271b0135" containerName="extract-content" Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: E1205 11:35:58.713725 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerName="extract-utilities" Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: I1205 11:35:58.713734 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerName="extract-utilities" Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: E1205 11:35:58.713764 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerName="extract-content" Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: I1205 11:35:58.713772 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerName="extract-content" Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: E1205 11:35:58.713788 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerName="registry-server" Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: I1205 11:35:58.713796 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerName="registry-server" Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: E1205 11:35:58.713829 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a41b09b-681e-423e-8b02-aa19271b0135" containerName="extract-utilities" Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: I1205 11:35:58.713836 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a41b09b-681e-423e-8b02-aa19271b0135" containerName="extract-utilities" Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: E1205 11:35:58.713852 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a41b09b-681e-423e-8b02-aa19271b0135" containerName="registry-server" Dec 05 11:35:58.714067 master-0 kubenswrapper[24928]: I1205 11:35:58.713858 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a41b09b-681e-423e-8b02-aa19271b0135" containerName="registry-server" Dec 05 11:35:58.714884 master-0 kubenswrapper[24928]: I1205 11:35:58.714173 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerName="registry-server" Dec 05 11:35:58.714884 master-0 kubenswrapper[24928]: I1205 11:35:58.714200 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerName="extract-content" Dec 05 11:35:58.714884 master-0 kubenswrapper[24928]: I1205 11:35:58.714225 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a41b09b-681e-423e-8b02-aa19271b0135" containerName="registry-server" Dec 05 11:35:58.714884 master-0 kubenswrapper[24928]: I1205 11:35:58.714264 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="0aa6e138-4b1f-4047-8255-a2b14d044588" containerName="extract-utilities" Dec 05 11:35:58.717785 master-0 kubenswrapper[24928]: I1205 11:35:58.717714 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:35:58.732668 master-0 kubenswrapper[24928]: I1205 11:35:58.732496 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4mvw4"] Dec 05 11:35:58.850773 master-0 kubenswrapper[24928]: I1205 11:35:58.850699 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-utilities\") pod \"certified-operators-4mvw4\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:35:58.851005 master-0 kubenswrapper[24928]: I1205 11:35:58.850810 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r46rg\" (UniqueName: \"kubernetes.io/projected/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-kube-api-access-r46rg\") pod \"certified-operators-4mvw4\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:35:58.851054 master-0 kubenswrapper[24928]: I1205 11:35:58.850992 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-catalog-content\") pod \"certified-operators-4mvw4\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:35:58.953330 master-0 kubenswrapper[24928]: I1205 11:35:58.953261 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r46rg\" (UniqueName: \"kubernetes.io/projected/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-kube-api-access-r46rg\") pod \"certified-operators-4mvw4\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:35:58.953586 master-0 kubenswrapper[24928]: I1205 11:35:58.953393 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-catalog-content\") pod \"certified-operators-4mvw4\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:35:58.953586 master-0 kubenswrapper[24928]: I1205 11:35:58.953580 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-utilities\") pod \"certified-operators-4mvw4\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:35:58.954129 master-0 kubenswrapper[24928]: I1205 11:35:58.954095 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-utilities\") pod \"certified-operators-4mvw4\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:35:58.954253 master-0 kubenswrapper[24928]: I1205 11:35:58.954208 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-catalog-content\") pod \"certified-operators-4mvw4\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:35:58.976954 master-0 kubenswrapper[24928]: I1205 11:35:58.976819 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r46rg\" (UniqueName: \"kubernetes.io/projected/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-kube-api-access-r46rg\") pod \"certified-operators-4mvw4\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:35:59.059107 master-0 kubenswrapper[24928]: I1205 11:35:59.059012 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:35:59.625786 master-0 kubenswrapper[24928]: I1205 11:35:59.625235 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4mvw4"] Dec 05 11:35:59.639872 master-0 kubenswrapper[24928]: W1205 11:35:59.637200 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2bcc2010_88ab_448e_b261_7e1e7e9b0f87.slice/crio-91dbc269b4179ad1861102e794219197a5f343119005e5dd24afdb76a31bcade WatchSource:0}: Error finding container 91dbc269b4179ad1861102e794219197a5f343119005e5dd24afdb76a31bcade: Status 404 returned error can't find the container with id 91dbc269b4179ad1861102e794219197a5f343119005e5dd24afdb76a31bcade Dec 05 11:36:00.487018 master-0 kubenswrapper[24928]: I1205 11:36:00.486960 24928 generic.go:334] "Generic (PLEG): container finished" podID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" containerID="f85c616595e79293a77b9ee858e49510be0c9fcbcb456ee38abcba15f0792d27" exitCode=0 Dec 05 11:36:00.487663 master-0 kubenswrapper[24928]: I1205 11:36:00.487608 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mvw4" event={"ID":"2bcc2010-88ab-448e-b261-7e1e7e9b0f87","Type":"ContainerDied","Data":"f85c616595e79293a77b9ee858e49510be0c9fcbcb456ee38abcba15f0792d27"} Dec 05 11:36:00.487790 master-0 kubenswrapper[24928]: I1205 11:36:00.487770 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mvw4" event={"ID":"2bcc2010-88ab-448e-b261-7e1e7e9b0f87","Type":"ContainerStarted","Data":"91dbc269b4179ad1861102e794219197a5f343119005e5dd24afdb76a31bcade"} Dec 05 11:36:01.505465 master-0 kubenswrapper[24928]: I1205 11:36:01.505276 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mvw4" event={"ID":"2bcc2010-88ab-448e-b261-7e1e7e9b0f87","Type":"ContainerStarted","Data":"1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e"} Dec 05 11:36:02.518971 master-0 kubenswrapper[24928]: I1205 11:36:02.518914 24928 generic.go:334] "Generic (PLEG): container finished" podID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" containerID="1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e" exitCode=0 Dec 05 11:36:02.519525 master-0 kubenswrapper[24928]: I1205 11:36:02.518980 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mvw4" event={"ID":"2bcc2010-88ab-448e-b261-7e1e7e9b0f87","Type":"ContainerDied","Data":"1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e"} Dec 05 11:36:03.536944 master-0 kubenswrapper[24928]: I1205 11:36:03.536793 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mvw4" event={"ID":"2bcc2010-88ab-448e-b261-7e1e7e9b0f87","Type":"ContainerStarted","Data":"00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27"} Dec 05 11:36:03.563990 master-0 kubenswrapper[24928]: I1205 11:36:03.563883 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4mvw4" podStartSLOduration=3.124806216 podStartE2EDuration="5.563865075s" podCreationTimestamp="2025-12-05 11:35:58 +0000 UTC" firstStartedPulling="2025-12-05 11:36:00.488906415 +0000 UTC m=+2920.492100266" lastFinishedPulling="2025-12-05 11:36:02.927965274 +0000 UTC m=+2922.931159125" observedRunningTime="2025-12-05 11:36:03.56036634 +0000 UTC m=+2923.563560191" watchObservedRunningTime="2025-12-05 11:36:03.563865075 +0000 UTC m=+2923.567058926" Dec 05 11:36:09.060979 master-0 kubenswrapper[24928]: I1205 11:36:09.060899 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:36:09.060979 master-0 kubenswrapper[24928]: I1205 11:36:09.060954 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:36:09.138418 master-0 kubenswrapper[24928]: I1205 11:36:09.138336 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:36:09.691851 master-0 kubenswrapper[24928]: I1205 11:36:09.691789 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:36:09.755602 master-0 kubenswrapper[24928]: I1205 11:36:09.755472 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4mvw4"] Dec 05 11:36:11.637777 master-0 kubenswrapper[24928]: I1205 11:36:11.637667 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4mvw4" podUID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" containerName="registry-server" containerID="cri-o://00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27" gracePeriod=2 Dec 05 11:36:12.266926 master-0 kubenswrapper[24928]: I1205 11:36:12.266839 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:36:12.326072 master-0 kubenswrapper[24928]: I1205 11:36:12.325962 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-catalog-content\") pod \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " Dec 05 11:36:12.326072 master-0 kubenswrapper[24928]: I1205 11:36:12.326014 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-utilities\") pod \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " Dec 05 11:36:12.326072 master-0 kubenswrapper[24928]: I1205 11:36:12.326060 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r46rg\" (UniqueName: \"kubernetes.io/projected/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-kube-api-access-r46rg\") pod \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\" (UID: \"2bcc2010-88ab-448e-b261-7e1e7e9b0f87\") " Dec 05 11:36:12.328394 master-0 kubenswrapper[24928]: I1205 11:36:12.328346 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-utilities" (OuterVolumeSpecName: "utilities") pod "2bcc2010-88ab-448e-b261-7e1e7e9b0f87" (UID: "2bcc2010-88ab-448e-b261-7e1e7e9b0f87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:36:12.333391 master-0 kubenswrapper[24928]: I1205 11:36:12.333320 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-kube-api-access-r46rg" (OuterVolumeSpecName: "kube-api-access-r46rg") pod "2bcc2010-88ab-448e-b261-7e1e7e9b0f87" (UID: "2bcc2010-88ab-448e-b261-7e1e7e9b0f87"). InnerVolumeSpecName "kube-api-access-r46rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:36:12.383184 master-0 kubenswrapper[24928]: I1205 11:36:12.383108 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2bcc2010-88ab-448e-b261-7e1e7e9b0f87" (UID: "2bcc2010-88ab-448e-b261-7e1e7e9b0f87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:36:12.430152 master-0 kubenswrapper[24928]: I1205 11:36:12.430013 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:36:12.430152 master-0 kubenswrapper[24928]: I1205 11:36:12.430065 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:36:12.430152 master-0 kubenswrapper[24928]: I1205 11:36:12.430081 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r46rg\" (UniqueName: \"kubernetes.io/projected/2bcc2010-88ab-448e-b261-7e1e7e9b0f87-kube-api-access-r46rg\") on node \"master-0\" DevicePath \"\"" Dec 05 11:36:12.657182 master-0 kubenswrapper[24928]: I1205 11:36:12.657096 24928 generic.go:334] "Generic (PLEG): container finished" podID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" containerID="00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27" exitCode=0 Dec 05 11:36:12.657182 master-0 kubenswrapper[24928]: I1205 11:36:12.657145 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mvw4" event={"ID":"2bcc2010-88ab-448e-b261-7e1e7e9b0f87","Type":"ContainerDied","Data":"00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27"} Dec 05 11:36:12.657182 master-0 kubenswrapper[24928]: I1205 11:36:12.657185 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4mvw4" event={"ID":"2bcc2010-88ab-448e-b261-7e1e7e9b0f87","Type":"ContainerDied","Data":"91dbc269b4179ad1861102e794219197a5f343119005e5dd24afdb76a31bcade"} Dec 05 11:36:12.657182 master-0 kubenswrapper[24928]: I1205 11:36:12.657205 24928 scope.go:117] "RemoveContainer" containerID="00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27" Dec 05 11:36:12.658617 master-0 kubenswrapper[24928]: I1205 11:36:12.657346 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4mvw4" Dec 05 11:36:12.684472 master-0 kubenswrapper[24928]: I1205 11:36:12.684361 24928 scope.go:117] "RemoveContainer" containerID="1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e" Dec 05 11:36:12.742726 master-0 kubenswrapper[24928]: I1205 11:36:12.742468 24928 scope.go:117] "RemoveContainer" containerID="f85c616595e79293a77b9ee858e49510be0c9fcbcb456ee38abcba15f0792d27" Dec 05 11:36:12.742726 master-0 kubenswrapper[24928]: I1205 11:36:12.742469 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4mvw4"] Dec 05 11:36:12.755303 master-0 kubenswrapper[24928]: I1205 11:36:12.755245 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4mvw4"] Dec 05 11:36:12.787725 master-0 kubenswrapper[24928]: I1205 11:36:12.787688 24928 scope.go:117] "RemoveContainer" containerID="00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27" Dec 05 11:36:12.788151 master-0 kubenswrapper[24928]: E1205 11:36:12.788041 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27\": container with ID starting with 00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27 not found: ID does not exist" containerID="00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27" Dec 05 11:36:12.788226 master-0 kubenswrapper[24928]: I1205 11:36:12.788150 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27"} err="failed to get container status \"00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27\": rpc error: code = NotFound desc = could not find container \"00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27\": container with ID starting with 00ada833d46845bc19c7ed36467985d5f2af13a8ce5969f35e4123acd3308b27 not found: ID does not exist" Dec 05 11:36:12.788226 master-0 kubenswrapper[24928]: I1205 11:36:12.788179 24928 scope.go:117] "RemoveContainer" containerID="1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e" Dec 05 11:36:12.788419 master-0 kubenswrapper[24928]: E1205 11:36:12.788378 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e\": container with ID starting with 1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e not found: ID does not exist" containerID="1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e" Dec 05 11:36:12.788508 master-0 kubenswrapper[24928]: I1205 11:36:12.788417 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e"} err="failed to get container status \"1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e\": rpc error: code = NotFound desc = could not find container \"1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e\": container with ID starting with 1efe370f20ef3dfef0a15d12f10da4b7a1a22c46226666e6de5c318c85a9e29e not found: ID does not exist" Dec 05 11:36:12.788508 master-0 kubenswrapper[24928]: I1205 11:36:12.788467 24928 scope.go:117] "RemoveContainer" containerID="f85c616595e79293a77b9ee858e49510be0c9fcbcb456ee38abcba15f0792d27" Dec 05 11:36:12.788709 master-0 kubenswrapper[24928]: E1205 11:36:12.788681 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f85c616595e79293a77b9ee858e49510be0c9fcbcb456ee38abcba15f0792d27\": container with ID starting with f85c616595e79293a77b9ee858e49510be0c9fcbcb456ee38abcba15f0792d27 not found: ID does not exist" containerID="f85c616595e79293a77b9ee858e49510be0c9fcbcb456ee38abcba15f0792d27" Dec 05 11:36:12.788779 master-0 kubenswrapper[24928]: I1205 11:36:12.788713 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f85c616595e79293a77b9ee858e49510be0c9fcbcb456ee38abcba15f0792d27"} err="failed to get container status \"f85c616595e79293a77b9ee858e49510be0c9fcbcb456ee38abcba15f0792d27\": rpc error: code = NotFound desc = could not find container \"f85c616595e79293a77b9ee858e49510be0c9fcbcb456ee38abcba15f0792d27\": container with ID starting with f85c616595e79293a77b9ee858e49510be0c9fcbcb456ee38abcba15f0792d27 not found: ID does not exist" Dec 05 11:36:14.218802 master-0 kubenswrapper[24928]: I1205 11:36:14.218728 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" path="/var/lib/kubelet/pods/2bcc2010-88ab-448e-b261-7e1e7e9b0f87/volumes" Dec 05 11:39:17.398091 master-0 kubenswrapper[24928]: E1205 11:39:17.398004 24928 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 192.168.32.10:50014->192.168.32.10:40581: write tcp 192.168.32.10:50014->192.168.32.10:40581: write: broken pipe Dec 05 11:42:48.471083 master-0 kubenswrapper[24928]: I1205 11:42:48.471009 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f5q2v"] Dec 05 11:42:48.472516 master-0 kubenswrapper[24928]: E1205 11:42:48.471568 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" containerName="extract-utilities" Dec 05 11:42:48.472516 master-0 kubenswrapper[24928]: I1205 11:42:48.471588 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" containerName="extract-utilities" Dec 05 11:42:48.472516 master-0 kubenswrapper[24928]: E1205 11:42:48.471611 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" containerName="registry-server" Dec 05 11:42:48.472516 master-0 kubenswrapper[24928]: I1205 11:42:48.471617 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" containerName="registry-server" Dec 05 11:42:48.472516 master-0 kubenswrapper[24928]: E1205 11:42:48.471643 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" containerName="extract-content" Dec 05 11:42:48.472516 master-0 kubenswrapper[24928]: I1205 11:42:48.471650 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" containerName="extract-content" Dec 05 11:42:48.472516 master-0 kubenswrapper[24928]: I1205 11:42:48.471928 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bcc2010-88ab-448e-b261-7e1e7e9b0f87" containerName="registry-server" Dec 05 11:42:48.473810 master-0 kubenswrapper[24928]: I1205 11:42:48.473776 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:48.497463 master-0 kubenswrapper[24928]: I1205 11:42:48.497374 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-catalog-content\") pod \"community-operators-f5q2v\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:48.497960 master-0 kubenswrapper[24928]: I1205 11:42:48.497911 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-utilities\") pod \"community-operators-f5q2v\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:48.498407 master-0 kubenswrapper[24928]: I1205 11:42:48.498370 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzxdx\" (UniqueName: \"kubernetes.io/projected/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-kube-api-access-pzxdx\") pod \"community-operators-f5q2v\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:48.525382 master-0 kubenswrapper[24928]: I1205 11:42:48.525316 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5q2v"] Dec 05 11:42:48.602001 master-0 kubenswrapper[24928]: I1205 11:42:48.601912 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-utilities\") pod \"community-operators-f5q2v\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:48.602304 master-0 kubenswrapper[24928]: I1205 11:42:48.602080 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzxdx\" (UniqueName: \"kubernetes.io/projected/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-kube-api-access-pzxdx\") pod \"community-operators-f5q2v\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:48.602304 master-0 kubenswrapper[24928]: I1205 11:42:48.602289 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-catalog-content\") pod \"community-operators-f5q2v\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:48.603381 master-0 kubenswrapper[24928]: I1205 11:42:48.603340 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-utilities\") pod \"community-operators-f5q2v\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:48.603574 master-0 kubenswrapper[24928]: I1205 11:42:48.603366 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-catalog-content\") pod \"community-operators-f5q2v\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:48.634492 master-0 kubenswrapper[24928]: I1205 11:42:48.631636 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzxdx\" (UniqueName: \"kubernetes.io/projected/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-kube-api-access-pzxdx\") pod \"community-operators-f5q2v\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:48.803130 master-0 kubenswrapper[24928]: I1205 11:42:48.803064 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:49.473538 master-0 kubenswrapper[24928]: W1205 11:42:49.468463 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97f5ed1a_123e_4e80_88ac_f4af8bd84b95.slice/crio-179d4cd3c6126a5e3212847a913cd08289898f7d62d753722a4799abce739c1e WatchSource:0}: Error finding container 179d4cd3c6126a5e3212847a913cd08289898f7d62d753722a4799abce739c1e: Status 404 returned error can't find the container with id 179d4cd3c6126a5e3212847a913cd08289898f7d62d753722a4799abce739c1e Dec 05 11:42:49.473538 master-0 kubenswrapper[24928]: I1205 11:42:49.469033 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f5q2v"] Dec 05 11:42:50.430528 master-0 kubenswrapper[24928]: I1205 11:42:50.430380 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5q2v" event={"ID":"97f5ed1a-123e-4e80-88ac-f4af8bd84b95","Type":"ContainerStarted","Data":"179d4cd3c6126a5e3212847a913cd08289898f7d62d753722a4799abce739c1e"} Dec 05 11:42:51.441252 master-0 kubenswrapper[24928]: I1205 11:42:51.441172 24928 generic.go:334] "Generic (PLEG): container finished" podID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" containerID="10f7b3f5fe578a72bca8933f667c818b98dc265cee8535af6deadd496558a9c6" exitCode=0 Dec 05 11:42:51.441825 master-0 kubenswrapper[24928]: I1205 11:42:51.441256 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5q2v" event={"ID":"97f5ed1a-123e-4e80-88ac-f4af8bd84b95","Type":"ContainerDied","Data":"10f7b3f5fe578a72bca8933f667c818b98dc265cee8535af6deadd496558a9c6"} Dec 05 11:42:51.444472 master-0 kubenswrapper[24928]: I1205 11:42:51.444413 24928 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:42:53.470879 master-0 kubenswrapper[24928]: I1205 11:42:53.470809 24928 generic.go:334] "Generic (PLEG): container finished" podID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" containerID="ecfca0728cb5a5f062eab05b8bd7152da02b18ceee1a22cbef74d95062ce3ee1" exitCode=0 Dec 05 11:42:53.470879 master-0 kubenswrapper[24928]: I1205 11:42:53.470853 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5q2v" event={"ID":"97f5ed1a-123e-4e80-88ac-f4af8bd84b95","Type":"ContainerDied","Data":"ecfca0728cb5a5f062eab05b8bd7152da02b18ceee1a22cbef74d95062ce3ee1"} Dec 05 11:42:54.485490 master-0 kubenswrapper[24928]: I1205 11:42:54.485406 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5q2v" event={"ID":"97f5ed1a-123e-4e80-88ac-f4af8bd84b95","Type":"ContainerStarted","Data":"4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750"} Dec 05 11:42:55.253461 master-0 kubenswrapper[24928]: I1205 11:42:55.252104 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f5q2v" podStartSLOduration=4.746602702 podStartE2EDuration="7.252080475s" podCreationTimestamp="2025-12-05 11:42:48 +0000 UTC" firstStartedPulling="2025-12-05 11:42:51.444214744 +0000 UTC m=+3331.447408595" lastFinishedPulling="2025-12-05 11:42:53.949692517 +0000 UTC m=+3333.952886368" observedRunningTime="2025-12-05 11:42:55.197478612 +0000 UTC m=+3335.200672523" watchObservedRunningTime="2025-12-05 11:42:55.252080475 +0000 UTC m=+3335.255274326" Dec 05 11:42:58.803531 master-0 kubenswrapper[24928]: I1205 11:42:58.803465 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:58.803531 master-0 kubenswrapper[24928]: I1205 11:42:58.803525 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:58.853266 master-0 kubenswrapper[24928]: I1205 11:42:58.853191 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:59.591111 master-0 kubenswrapper[24928]: I1205 11:42:59.591017 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:42:59.718740 master-0 kubenswrapper[24928]: I1205 11:42:59.718563 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f5q2v"] Dec 05 11:43:01.559837 master-0 kubenswrapper[24928]: I1205 11:43:01.559749 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f5q2v" podUID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" containerName="registry-server" containerID="cri-o://4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750" gracePeriod=2 Dec 05 11:43:02.105667 master-0 kubenswrapper[24928]: I1205 11:43:02.105615 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:43:02.251328 master-0 kubenswrapper[24928]: I1205 11:43:02.251154 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-catalog-content\") pod \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " Dec 05 11:43:02.251328 master-0 kubenswrapper[24928]: I1205 11:43:02.251246 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-utilities\") pod \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " Dec 05 11:43:02.251623 master-0 kubenswrapper[24928]: I1205 11:43:02.251519 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzxdx\" (UniqueName: \"kubernetes.io/projected/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-kube-api-access-pzxdx\") pod \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\" (UID: \"97f5ed1a-123e-4e80-88ac-f4af8bd84b95\") " Dec 05 11:43:02.252541 master-0 kubenswrapper[24928]: I1205 11:43:02.252498 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-utilities" (OuterVolumeSpecName: "utilities") pod "97f5ed1a-123e-4e80-88ac-f4af8bd84b95" (UID: "97f5ed1a-123e-4e80-88ac-f4af8bd84b95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:43:02.257039 master-0 kubenswrapper[24928]: I1205 11:43:02.256909 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-kube-api-access-pzxdx" (OuterVolumeSpecName: "kube-api-access-pzxdx") pod "97f5ed1a-123e-4e80-88ac-f4af8bd84b95" (UID: "97f5ed1a-123e-4e80-88ac-f4af8bd84b95"). InnerVolumeSpecName "kube-api-access-pzxdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:43:02.315268 master-0 kubenswrapper[24928]: I1205 11:43:02.315207 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97f5ed1a-123e-4e80-88ac-f4af8bd84b95" (UID: "97f5ed1a-123e-4e80-88ac-f4af8bd84b95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:43:02.354356 master-0 kubenswrapper[24928]: I1205 11:43:02.354281 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzxdx\" (UniqueName: \"kubernetes.io/projected/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-kube-api-access-pzxdx\") on node \"master-0\" DevicePath \"\"" Dec 05 11:43:02.354356 master-0 kubenswrapper[24928]: I1205 11:43:02.354347 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:43:02.354356 master-0 kubenswrapper[24928]: I1205 11:43:02.354359 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97f5ed1a-123e-4e80-88ac-f4af8bd84b95-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:43:02.574043 master-0 kubenswrapper[24928]: I1205 11:43:02.573990 24928 generic.go:334] "Generic (PLEG): container finished" podID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" containerID="4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750" exitCode=0 Dec 05 11:43:02.574890 master-0 kubenswrapper[24928]: I1205 11:43:02.574204 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5q2v" event={"ID":"97f5ed1a-123e-4e80-88ac-f4af8bd84b95","Type":"ContainerDied","Data":"4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750"} Dec 05 11:43:02.575000 master-0 kubenswrapper[24928]: I1205 11:43:02.574984 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f5q2v" event={"ID":"97f5ed1a-123e-4e80-88ac-f4af8bd84b95","Type":"ContainerDied","Data":"179d4cd3c6126a5e3212847a913cd08289898f7d62d753722a4799abce739c1e"} Dec 05 11:43:02.575080 master-0 kubenswrapper[24928]: I1205 11:43:02.575068 24928 scope.go:117] "RemoveContainer" containerID="4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750" Dec 05 11:43:02.575243 master-0 kubenswrapper[24928]: I1205 11:43:02.574281 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f5q2v" Dec 05 11:43:03.218510 master-0 kubenswrapper[24928]: I1205 11:43:03.218436 24928 scope.go:117] "RemoveContainer" containerID="ecfca0728cb5a5f062eab05b8bd7152da02b18ceee1a22cbef74d95062ce3ee1" Dec 05 11:43:03.262306 master-0 kubenswrapper[24928]: I1205 11:43:03.262249 24928 scope.go:117] "RemoveContainer" containerID="10f7b3f5fe578a72bca8933f667c818b98dc265cee8535af6deadd496558a9c6" Dec 05 11:43:03.296902 master-0 kubenswrapper[24928]: I1205 11:43:03.296160 24928 scope.go:117] "RemoveContainer" containerID="4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750" Dec 05 11:43:03.302488 master-0 kubenswrapper[24928]: I1205 11:43:03.300181 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f5q2v"] Dec 05 11:43:03.311911 master-0 kubenswrapper[24928]: E1205 11:43:03.311851 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750\": container with ID starting with 4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750 not found: ID does not exist" containerID="4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750" Dec 05 11:43:03.312030 master-0 kubenswrapper[24928]: I1205 11:43:03.311966 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750"} err="failed to get container status \"4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750\": rpc error: code = NotFound desc = could not find container \"4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750\": container with ID starting with 4b26fa3afbd45ec617bff0a7a99b2c392957dbacd920ece059e225868794c750 not found: ID does not exist" Dec 05 11:43:03.312030 master-0 kubenswrapper[24928]: I1205 11:43:03.312000 24928 scope.go:117] "RemoveContainer" containerID="ecfca0728cb5a5f062eab05b8bd7152da02b18ceee1a22cbef74d95062ce3ee1" Dec 05 11:43:03.312461 master-0 kubenswrapper[24928]: E1205 11:43:03.312433 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecfca0728cb5a5f062eab05b8bd7152da02b18ceee1a22cbef74d95062ce3ee1\": container with ID starting with ecfca0728cb5a5f062eab05b8bd7152da02b18ceee1a22cbef74d95062ce3ee1 not found: ID does not exist" containerID="ecfca0728cb5a5f062eab05b8bd7152da02b18ceee1a22cbef74d95062ce3ee1" Dec 05 11:43:03.312461 master-0 kubenswrapper[24928]: I1205 11:43:03.312455 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecfca0728cb5a5f062eab05b8bd7152da02b18ceee1a22cbef74d95062ce3ee1"} err="failed to get container status \"ecfca0728cb5a5f062eab05b8bd7152da02b18ceee1a22cbef74d95062ce3ee1\": rpc error: code = NotFound desc = could not find container \"ecfca0728cb5a5f062eab05b8bd7152da02b18ceee1a22cbef74d95062ce3ee1\": container with ID starting with ecfca0728cb5a5f062eab05b8bd7152da02b18ceee1a22cbef74d95062ce3ee1 not found: ID does not exist" Dec 05 11:43:03.312552 master-0 kubenswrapper[24928]: I1205 11:43:03.312468 24928 scope.go:117] "RemoveContainer" containerID="10f7b3f5fe578a72bca8933f667c818b98dc265cee8535af6deadd496558a9c6" Dec 05 11:43:03.313924 master-0 kubenswrapper[24928]: E1205 11:43:03.313864 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10f7b3f5fe578a72bca8933f667c818b98dc265cee8535af6deadd496558a9c6\": container with ID starting with 10f7b3f5fe578a72bca8933f667c818b98dc265cee8535af6deadd496558a9c6 not found: ID does not exist" containerID="10f7b3f5fe578a72bca8933f667c818b98dc265cee8535af6deadd496558a9c6" Dec 05 11:43:03.313924 master-0 kubenswrapper[24928]: I1205 11:43:03.313914 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10f7b3f5fe578a72bca8933f667c818b98dc265cee8535af6deadd496558a9c6"} err="failed to get container status \"10f7b3f5fe578a72bca8933f667c818b98dc265cee8535af6deadd496558a9c6\": rpc error: code = NotFound desc = could not find container \"10f7b3f5fe578a72bca8933f667c818b98dc265cee8535af6deadd496558a9c6\": container with ID starting with 10f7b3f5fe578a72bca8933f667c818b98dc265cee8535af6deadd496558a9c6 not found: ID does not exist" Dec 05 11:43:03.314069 master-0 kubenswrapper[24928]: I1205 11:43:03.314026 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f5q2v"] Dec 05 11:43:04.222577 master-0 kubenswrapper[24928]: I1205 11:43:04.222464 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" path="/var/lib/kubelet/pods/97f5ed1a-123e-4e80-88ac-f4af8bd84b95/volumes" Dec 05 11:43:31.204904 master-0 kubenswrapper[24928]: I1205 11:43:31.204736 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z7hpq"] Dec 05 11:43:31.205808 master-0 kubenswrapper[24928]: E1205 11:43:31.205265 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" containerName="extract-utilities" Dec 05 11:43:31.205808 master-0 kubenswrapper[24928]: I1205 11:43:31.205286 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" containerName="extract-utilities" Dec 05 11:43:31.205808 master-0 kubenswrapper[24928]: E1205 11:43:31.205344 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" containerName="registry-server" Dec 05 11:43:31.205808 master-0 kubenswrapper[24928]: I1205 11:43:31.205357 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" containerName="registry-server" Dec 05 11:43:31.205808 master-0 kubenswrapper[24928]: E1205 11:43:31.205381 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" containerName="extract-content" Dec 05 11:43:31.205808 master-0 kubenswrapper[24928]: I1205 11:43:31.205390 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" containerName="extract-content" Dec 05 11:43:31.205808 master-0 kubenswrapper[24928]: I1205 11:43:31.205705 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="97f5ed1a-123e-4e80-88ac-f4af8bd84b95" containerName="registry-server" Dec 05 11:43:31.210659 master-0 kubenswrapper[24928]: I1205 11:43:31.207463 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:31.241598 master-0 kubenswrapper[24928]: I1205 11:43:31.241034 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z7hpq"] Dec 05 11:43:31.308276 master-0 kubenswrapper[24928]: I1205 11:43:31.307658 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-catalog-content\") pod \"redhat-operators-z7hpq\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:31.308276 master-0 kubenswrapper[24928]: I1205 11:43:31.307778 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2xs2\" (UniqueName: \"kubernetes.io/projected/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-kube-api-access-s2xs2\") pod \"redhat-operators-z7hpq\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:31.308276 master-0 kubenswrapper[24928]: I1205 11:43:31.307944 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-utilities\") pod \"redhat-operators-z7hpq\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:31.410516 master-0 kubenswrapper[24928]: I1205 11:43:31.410449 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-utilities\") pod \"redhat-operators-z7hpq\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:31.410719 master-0 kubenswrapper[24928]: I1205 11:43:31.410651 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-catalog-content\") pod \"redhat-operators-z7hpq\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:31.410814 master-0 kubenswrapper[24928]: I1205 11:43:31.410714 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2xs2\" (UniqueName: \"kubernetes.io/projected/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-kube-api-access-s2xs2\") pod \"redhat-operators-z7hpq\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:31.411171 master-0 kubenswrapper[24928]: I1205 11:43:31.411119 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-utilities\") pod \"redhat-operators-z7hpq\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:31.411329 master-0 kubenswrapper[24928]: I1205 11:43:31.411274 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-catalog-content\") pod \"redhat-operators-z7hpq\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:33.007388 master-0 kubenswrapper[24928]: I1205 11:43:33.007335 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2xs2\" (UniqueName: \"kubernetes.io/projected/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-kube-api-access-s2xs2\") pod \"redhat-operators-z7hpq\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:33.046998 master-0 kubenswrapper[24928]: I1205 11:43:33.046931 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:33.571913 master-0 kubenswrapper[24928]: I1205 11:43:33.571839 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z7hpq"] Dec 05 11:43:34.966048 master-0 kubenswrapper[24928]: I1205 11:43:34.965972 24928 generic.go:334] "Generic (PLEG): container finished" podID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerID="2fed141f04fdb7a341877ce00cc05dba37d4415a8fa9468ab887889901cf686d" exitCode=0 Dec 05 11:43:34.966048 master-0 kubenswrapper[24928]: I1205 11:43:34.966029 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7hpq" event={"ID":"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7","Type":"ContainerDied","Data":"2fed141f04fdb7a341877ce00cc05dba37d4415a8fa9468ab887889901cf686d"} Dec 05 11:43:34.966048 master-0 kubenswrapper[24928]: I1205 11:43:34.966056 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7hpq" event={"ID":"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7","Type":"ContainerStarted","Data":"2d0d646301e5c7c802a26b8c035ca69f2c798fc2706d86a3fe0feba14a7b4af5"} Dec 05 11:43:36.993264 master-0 kubenswrapper[24928]: I1205 11:43:36.993170 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7hpq" event={"ID":"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7","Type":"ContainerStarted","Data":"391c05f62e6651add7bdb39b5d090a1bcdba7aa30ee8bb67c8c9f93bfd1caa66"} Dec 05 11:43:38.004292 master-0 kubenswrapper[24928]: I1205 11:43:38.004210 24928 generic.go:334] "Generic (PLEG): container finished" podID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerID="391c05f62e6651add7bdb39b5d090a1bcdba7aa30ee8bb67c8c9f93bfd1caa66" exitCode=0 Dec 05 11:43:38.004292 master-0 kubenswrapper[24928]: I1205 11:43:38.004270 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7hpq" event={"ID":"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7","Type":"ContainerDied","Data":"391c05f62e6651add7bdb39b5d090a1bcdba7aa30ee8bb67c8c9f93bfd1caa66"} Dec 05 11:43:41.050616 master-0 kubenswrapper[24928]: I1205 11:43:41.050502 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7hpq" event={"ID":"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7","Type":"ContainerStarted","Data":"aefb3499faa00d1b20083b5044768aa766b9a9d86c419a3623506aa8e6eaf39e"} Dec 05 11:43:41.076660 master-0 kubenswrapper[24928]: I1205 11:43:41.076553 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z7hpq" podStartSLOduration=4.918075925 podStartE2EDuration="10.076531214s" podCreationTimestamp="2025-12-05 11:43:31 +0000 UTC" firstStartedPulling="2025-12-05 11:43:34.967747644 +0000 UTC m=+3374.970941495" lastFinishedPulling="2025-12-05 11:43:40.126202933 +0000 UTC m=+3380.129396784" observedRunningTime="2025-12-05 11:43:41.07232722 +0000 UTC m=+3381.075521071" watchObservedRunningTime="2025-12-05 11:43:41.076531214 +0000 UTC m=+3381.079725065" Dec 05 11:43:43.048050 master-0 kubenswrapper[24928]: I1205 11:43:43.047637 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:43.048050 master-0 kubenswrapper[24928]: I1205 11:43:43.047821 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:44.111323 master-0 kubenswrapper[24928]: I1205 11:43:44.111219 24928 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z7hpq" podUID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerName="registry-server" probeResult="failure" output=< Dec 05 11:43:44.111323 master-0 kubenswrapper[24928]: timeout: failed to connect service ":50051" within 1s Dec 05 11:43:44.111323 master-0 kubenswrapper[24928]: > Dec 05 11:43:53.098764 master-0 kubenswrapper[24928]: I1205 11:43:53.098546 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:53.150919 master-0 kubenswrapper[24928]: I1205 11:43:53.150842 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:55.044252 master-0 kubenswrapper[24928]: I1205 11:43:55.044123 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z7hpq"] Dec 05 11:43:55.045169 master-0 kubenswrapper[24928]: I1205 11:43:55.045094 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z7hpq" podUID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerName="registry-server" containerID="cri-o://aefb3499faa00d1b20083b5044768aa766b9a9d86c419a3623506aa8e6eaf39e" gracePeriod=2 Dec 05 11:43:55.234451 master-0 kubenswrapper[24928]: I1205 11:43:55.234361 24928 generic.go:334] "Generic (PLEG): container finished" podID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerID="aefb3499faa00d1b20083b5044768aa766b9a9d86c419a3623506aa8e6eaf39e" exitCode=0 Dec 05 11:43:55.234451 master-0 kubenswrapper[24928]: I1205 11:43:55.234414 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7hpq" event={"ID":"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7","Type":"ContainerDied","Data":"aefb3499faa00d1b20083b5044768aa766b9a9d86c419a3623506aa8e6eaf39e"} Dec 05 11:43:55.800323 master-0 kubenswrapper[24928]: I1205 11:43:55.800230 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:55.933113 master-0 kubenswrapper[24928]: I1205 11:43:55.933018 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2xs2\" (UniqueName: \"kubernetes.io/projected/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-kube-api-access-s2xs2\") pod \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " Dec 05 11:43:55.933113 master-0 kubenswrapper[24928]: I1205 11:43:55.933105 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-catalog-content\") pod \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " Dec 05 11:43:55.933601 master-0 kubenswrapper[24928]: I1205 11:43:55.933197 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-utilities\") pod \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\" (UID: \"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7\") " Dec 05 11:43:55.935015 master-0 kubenswrapper[24928]: I1205 11:43:55.934865 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-utilities" (OuterVolumeSpecName: "utilities") pod "ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" (UID: "ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:43:55.937206 master-0 kubenswrapper[24928]: I1205 11:43:55.937134 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-kube-api-access-s2xs2" (OuterVolumeSpecName: "kube-api-access-s2xs2") pod "ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" (UID: "ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7"). InnerVolumeSpecName "kube-api-access-s2xs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:43:56.038149 master-0 kubenswrapper[24928]: I1205 11:43:56.038026 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2xs2\" (UniqueName: \"kubernetes.io/projected/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-kube-api-access-s2xs2\") on node \"master-0\" DevicePath \"\"" Dec 05 11:43:56.038149 master-0 kubenswrapper[24928]: I1205 11:43:56.038113 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:43:56.044498 master-0 kubenswrapper[24928]: I1205 11:43:56.044376 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" (UID: "ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:43:56.140955 master-0 kubenswrapper[24928]: I1205 11:43:56.140857 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:43:56.253788 master-0 kubenswrapper[24928]: I1205 11:43:56.253592 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z7hpq" event={"ID":"ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7","Type":"ContainerDied","Data":"2d0d646301e5c7c802a26b8c035ca69f2c798fc2706d86a3fe0feba14a7b4af5"} Dec 05 11:43:56.253788 master-0 kubenswrapper[24928]: I1205 11:43:56.253668 24928 scope.go:117] "RemoveContainer" containerID="aefb3499faa00d1b20083b5044768aa766b9a9d86c419a3623506aa8e6eaf39e" Dec 05 11:43:56.253788 master-0 kubenswrapper[24928]: I1205 11:43:56.253719 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z7hpq" Dec 05 11:43:56.280619 master-0 kubenswrapper[24928]: I1205 11:43:56.279798 24928 scope.go:117] "RemoveContainer" containerID="391c05f62e6651add7bdb39b5d090a1bcdba7aa30ee8bb67c8c9f93bfd1caa66" Dec 05 11:43:56.295587 master-0 kubenswrapper[24928]: I1205 11:43:56.295460 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z7hpq"] Dec 05 11:43:56.309559 master-0 kubenswrapper[24928]: I1205 11:43:56.309464 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z7hpq"] Dec 05 11:43:56.683404 master-0 kubenswrapper[24928]: I1205 11:43:56.683336 24928 scope.go:117] "RemoveContainer" containerID="2fed141f04fdb7a341877ce00cc05dba37d4415a8fa9468ab887889901cf686d" Dec 05 11:43:58.220815 master-0 kubenswrapper[24928]: I1205 11:43:58.220740 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" path="/var/lib/kubelet/pods/ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7/volumes" Dec 05 11:44:09.142143 master-0 kubenswrapper[24928]: I1205 11:44:09.142091 24928 trace.go:236] Trace[1255302637]: "Calculate volume metrics of var-lib-ironic for pod openstack/ironic-conductor-0" (05-Dec-2025 11:44:08.027) (total time: 1114ms): Dec 05 11:44:09.142143 master-0 kubenswrapper[24928]: Trace[1255302637]: [1.114092079s] [1.114092079s] END Dec 05 11:44:23.815059 master-0 kubenswrapper[24928]: I1205 11:44:23.814996 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d8t88"] Dec 05 11:44:23.815815 master-0 kubenswrapper[24928]: E1205 11:44:23.815609 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerName="registry-server" Dec 05 11:44:23.815815 master-0 kubenswrapper[24928]: I1205 11:44:23.815625 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerName="registry-server" Dec 05 11:44:23.815815 master-0 kubenswrapper[24928]: E1205 11:44:23.815647 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerName="extract-utilities" Dec 05 11:44:23.815815 master-0 kubenswrapper[24928]: I1205 11:44:23.815655 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerName="extract-utilities" Dec 05 11:44:23.815815 master-0 kubenswrapper[24928]: E1205 11:44:23.815665 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerName="extract-content" Dec 05 11:44:23.815815 master-0 kubenswrapper[24928]: I1205 11:44:23.815671 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerName="extract-content" Dec 05 11:44:23.816067 master-0 kubenswrapper[24928]: I1205 11:44:23.815952 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce82ebe0-5ae0-4f18-80f6-748c9a31a8e7" containerName="registry-server" Dec 05 11:44:23.817995 master-0 kubenswrapper[24928]: I1205 11:44:23.817962 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:23.831539 master-0 kubenswrapper[24928]: I1205 11:44:23.831476 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d8t88"] Dec 05 11:44:23.907441 master-0 kubenswrapper[24928]: I1205 11:44:23.907371 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9d6z\" (UniqueName: \"kubernetes.io/projected/cc1e8fec-65f6-41bf-947d-7bc905ecee56-kube-api-access-x9d6z\") pod \"redhat-marketplace-d8t88\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:23.908257 master-0 kubenswrapper[24928]: I1205 11:44:23.908207 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-utilities\") pod \"redhat-marketplace-d8t88\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:23.909748 master-0 kubenswrapper[24928]: I1205 11:44:23.908651 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-catalog-content\") pod \"redhat-marketplace-d8t88\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:24.011312 master-0 kubenswrapper[24928]: I1205 11:44:24.011252 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-catalog-content\") pod \"redhat-marketplace-d8t88\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:24.011408 master-0 kubenswrapper[24928]: I1205 11:44:24.011395 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9d6z\" (UniqueName: \"kubernetes.io/projected/cc1e8fec-65f6-41bf-947d-7bc905ecee56-kube-api-access-x9d6z\") pod \"redhat-marketplace-d8t88\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:24.011497 master-0 kubenswrapper[24928]: I1205 11:44:24.011470 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-utilities\") pod \"redhat-marketplace-d8t88\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:24.011880 master-0 kubenswrapper[24928]: I1205 11:44:24.011836 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-catalog-content\") pod \"redhat-marketplace-d8t88\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:24.011880 master-0 kubenswrapper[24928]: I1205 11:44:24.011867 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-utilities\") pod \"redhat-marketplace-d8t88\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:24.029276 master-0 kubenswrapper[24928]: I1205 11:44:24.027799 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9d6z\" (UniqueName: \"kubernetes.io/projected/cc1e8fec-65f6-41bf-947d-7bc905ecee56-kube-api-access-x9d6z\") pod \"redhat-marketplace-d8t88\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:24.144354 master-0 kubenswrapper[24928]: I1205 11:44:24.144280 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:24.631065 master-0 kubenswrapper[24928]: I1205 11:44:24.631006 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d8t88"] Dec 05 11:44:25.619800 master-0 kubenswrapper[24928]: I1205 11:44:25.619704 24928 generic.go:334] "Generic (PLEG): container finished" podID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" containerID="bed154aab58f32ba7c51add6eee605345c5fdcad7ee95822e415585dc489347d" exitCode=0 Dec 05 11:44:25.620411 master-0 kubenswrapper[24928]: I1205 11:44:25.619985 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8t88" event={"ID":"cc1e8fec-65f6-41bf-947d-7bc905ecee56","Type":"ContainerDied","Data":"bed154aab58f32ba7c51add6eee605345c5fdcad7ee95822e415585dc489347d"} Dec 05 11:44:25.620411 master-0 kubenswrapper[24928]: I1205 11:44:25.620048 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8t88" event={"ID":"cc1e8fec-65f6-41bf-947d-7bc905ecee56","Type":"ContainerStarted","Data":"196d4391db603a97a077a5cbbdd2abc7aaca844c364c14e19309c8986c055a45"} Dec 05 11:44:27.643279 master-0 kubenswrapper[24928]: I1205 11:44:27.643191 24928 generic.go:334] "Generic (PLEG): container finished" podID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" containerID="0da35409fd65e16b269a5639e2d671770ed9011b711705de4a806ffaf4e40ed6" exitCode=0 Dec 05 11:44:27.643279 master-0 kubenswrapper[24928]: I1205 11:44:27.643263 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8t88" event={"ID":"cc1e8fec-65f6-41bf-947d-7bc905ecee56","Type":"ContainerDied","Data":"0da35409fd65e16b269a5639e2d671770ed9011b711705de4a806ffaf4e40ed6"} Dec 05 11:44:28.656001 master-0 kubenswrapper[24928]: I1205 11:44:28.655859 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8t88" event={"ID":"cc1e8fec-65f6-41bf-947d-7bc905ecee56","Type":"ContainerStarted","Data":"04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6"} Dec 05 11:44:28.700128 master-0 kubenswrapper[24928]: I1205 11:44:28.699239 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d8t88" podStartSLOduration=3.230778937 podStartE2EDuration="5.699217739s" podCreationTimestamp="2025-12-05 11:44:23 +0000 UTC" firstStartedPulling="2025-12-05 11:44:25.622397547 +0000 UTC m=+3425.625591398" lastFinishedPulling="2025-12-05 11:44:28.090836339 +0000 UTC m=+3428.094030200" observedRunningTime="2025-12-05 11:44:28.686660897 +0000 UTC m=+3428.689854758" watchObservedRunningTime="2025-12-05 11:44:28.699217739 +0000 UTC m=+3428.702411590" Dec 05 11:44:34.145408 master-0 kubenswrapper[24928]: I1205 11:44:34.145323 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:34.145408 master-0 kubenswrapper[24928]: I1205 11:44:34.145398 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:34.196575 master-0 kubenswrapper[24928]: I1205 11:44:34.196477 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:34.807206 master-0 kubenswrapper[24928]: I1205 11:44:34.807139 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:34.873205 master-0 kubenswrapper[24928]: I1205 11:44:34.873142 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d8t88"] Dec 05 11:44:36.768153 master-0 kubenswrapper[24928]: I1205 11:44:36.767241 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d8t88" podUID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" containerName="registry-server" containerID="cri-o://04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6" gracePeriod=2 Dec 05 11:44:37.326824 master-0 kubenswrapper[24928]: I1205 11:44:37.326758 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:37.467207 master-0 kubenswrapper[24928]: I1205 11:44:37.467128 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9d6z\" (UniqueName: \"kubernetes.io/projected/cc1e8fec-65f6-41bf-947d-7bc905ecee56-kube-api-access-x9d6z\") pod \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " Dec 05 11:44:37.467599 master-0 kubenswrapper[24928]: I1205 11:44:37.467574 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-utilities\") pod \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " Dec 05 11:44:37.467827 master-0 kubenswrapper[24928]: I1205 11:44:37.467806 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-catalog-content\") pod \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\" (UID: \"cc1e8fec-65f6-41bf-947d-7bc905ecee56\") " Dec 05 11:44:37.469625 master-0 kubenswrapper[24928]: I1205 11:44:37.469358 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-utilities" (OuterVolumeSpecName: "utilities") pod "cc1e8fec-65f6-41bf-947d-7bc905ecee56" (UID: "cc1e8fec-65f6-41bf-947d-7bc905ecee56"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:44:37.474541 master-0 kubenswrapper[24928]: I1205 11:44:37.474445 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc1e8fec-65f6-41bf-947d-7bc905ecee56-kube-api-access-x9d6z" (OuterVolumeSpecName: "kube-api-access-x9d6z") pod "cc1e8fec-65f6-41bf-947d-7bc905ecee56" (UID: "cc1e8fec-65f6-41bf-947d-7bc905ecee56"). InnerVolumeSpecName "kube-api-access-x9d6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:44:37.489528 master-0 kubenswrapper[24928]: I1205 11:44:37.488641 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc1e8fec-65f6-41bf-947d-7bc905ecee56" (UID: "cc1e8fec-65f6-41bf-947d-7bc905ecee56"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:44:37.571646 master-0 kubenswrapper[24928]: I1205 11:44:37.571584 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9d6z\" (UniqueName: \"kubernetes.io/projected/cc1e8fec-65f6-41bf-947d-7bc905ecee56-kube-api-access-x9d6z\") on node \"master-0\" DevicePath \"\"" Dec 05 11:44:37.571646 master-0 kubenswrapper[24928]: I1205 11:44:37.571629 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:44:37.571646 master-0 kubenswrapper[24928]: I1205 11:44:37.571638 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc1e8fec-65f6-41bf-947d-7bc905ecee56-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:44:37.789859 master-0 kubenswrapper[24928]: I1205 11:44:37.789774 24928 generic.go:334] "Generic (PLEG): container finished" podID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" containerID="04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6" exitCode=0 Dec 05 11:44:37.789859 master-0 kubenswrapper[24928]: I1205 11:44:37.789849 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8t88" event={"ID":"cc1e8fec-65f6-41bf-947d-7bc905ecee56","Type":"ContainerDied","Data":"04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6"} Dec 05 11:44:37.790601 master-0 kubenswrapper[24928]: I1205 11:44:37.789885 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d8t88" event={"ID":"cc1e8fec-65f6-41bf-947d-7bc905ecee56","Type":"ContainerDied","Data":"196d4391db603a97a077a5cbbdd2abc7aaca844c364c14e19309c8986c055a45"} Dec 05 11:44:37.790601 master-0 kubenswrapper[24928]: I1205 11:44:37.789907 24928 scope.go:117] "RemoveContainer" containerID="04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6" Dec 05 11:44:37.790601 master-0 kubenswrapper[24928]: I1205 11:44:37.789982 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d8t88" Dec 05 11:44:37.815212 master-0 kubenswrapper[24928]: I1205 11:44:37.815160 24928 scope.go:117] "RemoveContainer" containerID="0da35409fd65e16b269a5639e2d671770ed9011b711705de4a806ffaf4e40ed6" Dec 05 11:44:37.837149 master-0 kubenswrapper[24928]: I1205 11:44:37.837108 24928 scope.go:117] "RemoveContainer" containerID="bed154aab58f32ba7c51add6eee605345c5fdcad7ee95822e415585dc489347d" Dec 05 11:44:37.897933 master-0 kubenswrapper[24928]: I1205 11:44:37.897886 24928 scope.go:117] "RemoveContainer" containerID="04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6" Dec 05 11:44:37.898471 master-0 kubenswrapper[24928]: E1205 11:44:37.898390 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6\": container with ID starting with 04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6 not found: ID does not exist" containerID="04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6" Dec 05 11:44:37.898471 master-0 kubenswrapper[24928]: I1205 11:44:37.898445 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6"} err="failed to get container status \"04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6\": rpc error: code = NotFound desc = could not find container \"04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6\": container with ID starting with 04caf90e85cac322eb7cfd55f27c02e27208efcd05e4762ab79ab29227cf63c6 not found: ID does not exist" Dec 05 11:44:37.898471 master-0 kubenswrapper[24928]: I1205 11:44:37.898470 24928 scope.go:117] "RemoveContainer" containerID="0da35409fd65e16b269a5639e2d671770ed9011b711705de4a806ffaf4e40ed6" Dec 05 11:44:37.898883 master-0 kubenswrapper[24928]: E1205 11:44:37.898844 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0da35409fd65e16b269a5639e2d671770ed9011b711705de4a806ffaf4e40ed6\": container with ID starting with 0da35409fd65e16b269a5639e2d671770ed9011b711705de4a806ffaf4e40ed6 not found: ID does not exist" containerID="0da35409fd65e16b269a5639e2d671770ed9011b711705de4a806ffaf4e40ed6" Dec 05 11:44:37.898938 master-0 kubenswrapper[24928]: I1205 11:44:37.898883 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0da35409fd65e16b269a5639e2d671770ed9011b711705de4a806ffaf4e40ed6"} err="failed to get container status \"0da35409fd65e16b269a5639e2d671770ed9011b711705de4a806ffaf4e40ed6\": rpc error: code = NotFound desc = could not find container \"0da35409fd65e16b269a5639e2d671770ed9011b711705de4a806ffaf4e40ed6\": container with ID starting with 0da35409fd65e16b269a5639e2d671770ed9011b711705de4a806ffaf4e40ed6 not found: ID does not exist" Dec 05 11:44:37.898938 master-0 kubenswrapper[24928]: I1205 11:44:37.898902 24928 scope.go:117] "RemoveContainer" containerID="bed154aab58f32ba7c51add6eee605345c5fdcad7ee95822e415585dc489347d" Dec 05 11:44:37.899188 master-0 kubenswrapper[24928]: E1205 11:44:37.899158 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bed154aab58f32ba7c51add6eee605345c5fdcad7ee95822e415585dc489347d\": container with ID starting with bed154aab58f32ba7c51add6eee605345c5fdcad7ee95822e415585dc489347d not found: ID does not exist" containerID="bed154aab58f32ba7c51add6eee605345c5fdcad7ee95822e415585dc489347d" Dec 05 11:44:37.899188 master-0 kubenswrapper[24928]: I1205 11:44:37.899183 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bed154aab58f32ba7c51add6eee605345c5fdcad7ee95822e415585dc489347d"} err="failed to get container status \"bed154aab58f32ba7c51add6eee605345c5fdcad7ee95822e415585dc489347d\": rpc error: code = NotFound desc = could not find container \"bed154aab58f32ba7c51add6eee605345c5fdcad7ee95822e415585dc489347d\": container with ID starting with bed154aab58f32ba7c51add6eee605345c5fdcad7ee95822e415585dc489347d not found: ID does not exist" Dec 05 11:44:37.962391 master-0 kubenswrapper[24928]: I1205 11:44:37.960783 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d8t88"] Dec 05 11:44:37.976651 master-0 kubenswrapper[24928]: I1205 11:44:37.976571 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d8t88"] Dec 05 11:44:38.217084 master-0 kubenswrapper[24928]: I1205 11:44:38.217012 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" path="/var/lib/kubelet/pods/cc1e8fec-65f6-41bf-947d-7bc905ecee56/volumes" Dec 05 11:45:00.295467 master-0 kubenswrapper[24928]: I1205 11:45:00.295352 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27"] Dec 05 11:45:00.296232 master-0 kubenswrapper[24928]: E1205 11:45:00.295887 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" containerName="registry-server" Dec 05 11:45:00.296232 master-0 kubenswrapper[24928]: I1205 11:45:00.295903 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" containerName="registry-server" Dec 05 11:45:00.296232 master-0 kubenswrapper[24928]: E1205 11:45:00.295919 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" containerName="extract-content" Dec 05 11:45:00.296232 master-0 kubenswrapper[24928]: I1205 11:45:00.295925 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" containerName="extract-content" Dec 05 11:45:00.296232 master-0 kubenswrapper[24928]: E1205 11:45:00.295977 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" containerName="extract-utilities" Dec 05 11:45:00.296232 master-0 kubenswrapper[24928]: I1205 11:45:00.295984 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" containerName="extract-utilities" Dec 05 11:45:00.296232 master-0 kubenswrapper[24928]: I1205 11:45:00.296216 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc1e8fec-65f6-41bf-947d-7bc905ecee56" containerName="registry-server" Dec 05 11:45:00.297012 master-0 kubenswrapper[24928]: I1205 11:45:00.296964 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:00.299570 master-0 kubenswrapper[24928]: I1205 11:45:00.299376 24928 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-fdxl6" Dec 05 11:45:00.300693 master-0 kubenswrapper[24928]: I1205 11:45:00.300657 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 11:45:00.353624 master-0 kubenswrapper[24928]: I1205 11:45:00.353544 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27"] Dec 05 11:45:00.408714 master-0 kubenswrapper[24928]: I1205 11:45:00.408608 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clqbl\" (UniqueName: \"kubernetes.io/projected/f884d097-5af4-48b5-8a84-173dcaeedf4e-kube-api-access-clqbl\") pod \"collect-profiles-29415585-rjr27\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:00.408963 master-0 kubenswrapper[24928]: I1205 11:45:00.408743 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f884d097-5af4-48b5-8a84-173dcaeedf4e-secret-volume\") pod \"collect-profiles-29415585-rjr27\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:00.408963 master-0 kubenswrapper[24928]: I1205 11:45:00.408841 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f884d097-5af4-48b5-8a84-173dcaeedf4e-config-volume\") pod \"collect-profiles-29415585-rjr27\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:00.514691 master-0 kubenswrapper[24928]: I1205 11:45:00.514606 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f884d097-5af4-48b5-8a84-173dcaeedf4e-config-volume\") pod \"collect-profiles-29415585-rjr27\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:00.515276 master-0 kubenswrapper[24928]: I1205 11:45:00.515243 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clqbl\" (UniqueName: \"kubernetes.io/projected/f884d097-5af4-48b5-8a84-173dcaeedf4e-kube-api-access-clqbl\") pod \"collect-profiles-29415585-rjr27\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:00.515601 master-0 kubenswrapper[24928]: I1205 11:45:00.515578 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f884d097-5af4-48b5-8a84-173dcaeedf4e-secret-volume\") pod \"collect-profiles-29415585-rjr27\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:00.515788 master-0 kubenswrapper[24928]: I1205 11:45:00.515737 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f884d097-5af4-48b5-8a84-173dcaeedf4e-config-volume\") pod \"collect-profiles-29415585-rjr27\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:00.521554 master-0 kubenswrapper[24928]: I1205 11:45:00.521355 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f884d097-5af4-48b5-8a84-173dcaeedf4e-secret-volume\") pod \"collect-profiles-29415585-rjr27\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:00.532711 master-0 kubenswrapper[24928]: I1205 11:45:00.532303 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clqbl\" (UniqueName: \"kubernetes.io/projected/f884d097-5af4-48b5-8a84-173dcaeedf4e-kube-api-access-clqbl\") pod \"collect-profiles-29415585-rjr27\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:00.616855 master-0 kubenswrapper[24928]: I1205 11:45:00.616709 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:01.176034 master-0 kubenswrapper[24928]: I1205 11:45:01.175975 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27"] Dec 05 11:45:02.080585 master-0 kubenswrapper[24928]: I1205 11:45:02.080067 24928 generic.go:334] "Generic (PLEG): container finished" podID="f884d097-5af4-48b5-8a84-173dcaeedf4e" containerID="dba08532415242fd504d004dd92c975aa3bba629ccff37dc623b3282f65dd127" exitCode=0 Dec 05 11:45:02.080585 master-0 kubenswrapper[24928]: I1205 11:45:02.080150 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" event={"ID":"f884d097-5af4-48b5-8a84-173dcaeedf4e","Type":"ContainerDied","Data":"dba08532415242fd504d004dd92c975aa3bba629ccff37dc623b3282f65dd127"} Dec 05 11:45:02.080585 master-0 kubenswrapper[24928]: I1205 11:45:02.080215 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" event={"ID":"f884d097-5af4-48b5-8a84-173dcaeedf4e","Type":"ContainerStarted","Data":"d8a880ee45713a66d2a65540871503e4af2232d8688ed635ee8c3d60222209d1"} Dec 05 11:45:03.551083 master-0 kubenswrapper[24928]: I1205 11:45:03.551022 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:03.605858 master-0 kubenswrapper[24928]: I1205 11:45:03.605739 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f884d097-5af4-48b5-8a84-173dcaeedf4e-config-volume\") pod \"f884d097-5af4-48b5-8a84-173dcaeedf4e\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " Dec 05 11:45:03.605858 master-0 kubenswrapper[24928]: I1205 11:45:03.605847 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clqbl\" (UniqueName: \"kubernetes.io/projected/f884d097-5af4-48b5-8a84-173dcaeedf4e-kube-api-access-clqbl\") pod \"f884d097-5af4-48b5-8a84-173dcaeedf4e\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " Dec 05 11:45:03.606308 master-0 kubenswrapper[24928]: I1205 11:45:03.605955 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f884d097-5af4-48b5-8a84-173dcaeedf4e-secret-volume\") pod \"f884d097-5af4-48b5-8a84-173dcaeedf4e\" (UID: \"f884d097-5af4-48b5-8a84-173dcaeedf4e\") " Dec 05 11:45:03.606461 master-0 kubenswrapper[24928]: I1205 11:45:03.606380 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f884d097-5af4-48b5-8a84-173dcaeedf4e-config-volume" (OuterVolumeSpecName: "config-volume") pod "f884d097-5af4-48b5-8a84-173dcaeedf4e" (UID: "f884d097-5af4-48b5-8a84-173dcaeedf4e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 11:45:03.607396 master-0 kubenswrapper[24928]: I1205 11:45:03.607348 24928 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f884d097-5af4-48b5-8a84-173dcaeedf4e-config-volume\") on node \"master-0\" DevicePath \"\"" Dec 05 11:45:03.609131 master-0 kubenswrapper[24928]: I1205 11:45:03.609062 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f884d097-5af4-48b5-8a84-173dcaeedf4e-kube-api-access-clqbl" (OuterVolumeSpecName: "kube-api-access-clqbl") pod "f884d097-5af4-48b5-8a84-173dcaeedf4e" (UID: "f884d097-5af4-48b5-8a84-173dcaeedf4e"). InnerVolumeSpecName "kube-api-access-clqbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:45:03.609394 master-0 kubenswrapper[24928]: I1205 11:45:03.609323 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f884d097-5af4-48b5-8a84-173dcaeedf4e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f884d097-5af4-48b5-8a84-173dcaeedf4e" (UID: "f884d097-5af4-48b5-8a84-173dcaeedf4e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 11:45:03.710458 master-0 kubenswrapper[24928]: I1205 11:45:03.710235 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clqbl\" (UniqueName: \"kubernetes.io/projected/f884d097-5af4-48b5-8a84-173dcaeedf4e-kube-api-access-clqbl\") on node \"master-0\" DevicePath \"\"" Dec 05 11:45:03.710769 master-0 kubenswrapper[24928]: I1205 11:45:03.710750 24928 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f884d097-5af4-48b5-8a84-173dcaeedf4e-secret-volume\") on node \"master-0\" DevicePath \"\"" Dec 05 11:45:04.107214 master-0 kubenswrapper[24928]: I1205 11:45:04.107155 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" event={"ID":"f884d097-5af4-48b5-8a84-173dcaeedf4e","Type":"ContainerDied","Data":"d8a880ee45713a66d2a65540871503e4af2232d8688ed635ee8c3d60222209d1"} Dec 05 11:45:04.107608 master-0 kubenswrapper[24928]: I1205 11:45:04.107592 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8a880ee45713a66d2a65540871503e4af2232d8688ed635ee8c3d60222209d1" Dec 05 11:45:04.107690 master-0 kubenswrapper[24928]: I1205 11:45:04.107221 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415585-rjr27" Dec 05 11:45:04.253526 master-0 kubenswrapper[24928]: E1205 11:45:04.253461 24928 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf884d097_5af4_48b5_8a84_173dcaeedf4e.slice/crio-d8a880ee45713a66d2a65540871503e4af2232d8688ed635ee8c3d60222209d1\": RecentStats: unable to find data in memory cache]" Dec 05 11:45:04.924411 master-0 kubenswrapper[24928]: I1205 11:45:04.924341 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm"] Dec 05 11:45:04.934305 master-0 kubenswrapper[24928]: I1205 11:45:04.934240 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415540-dgqvm"] Dec 05 11:45:06.220971 master-0 kubenswrapper[24928]: I1205 11:45:06.220891 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06b902d6-85f1-4661-8cc6-be82f74a966d" path="/var/lib/kubelet/pods/06b902d6-85f1-4661-8cc6-be82f74a966d/volumes" Dec 05 11:45:49.755411 master-0 kubenswrapper[24928]: I1205 11:45:49.755242 24928 scope.go:117] "RemoveContainer" containerID="77c9b6fd5d013100eb68e4afeb5ea7a33da6bb0eb6942ac8e68f5e51cd472646" Dec 05 11:46:17.925941 master-0 kubenswrapper[24928]: E1205 11:46:17.925831 24928 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 192.168.32.10:44252->192.168.32.10:40581: write tcp 192.168.32.10:44252->192.168.32.10:40581: write: broken pipe Dec 05 11:46:19.245963 master-0 kubenswrapper[24928]: I1205 11:46:19.245883 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-crjkl"] Dec 05 11:46:19.246609 master-0 kubenswrapper[24928]: E1205 11:46:19.246585 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f884d097-5af4-48b5-8a84-173dcaeedf4e" containerName="collect-profiles" Dec 05 11:46:19.246609 master-0 kubenswrapper[24928]: I1205 11:46:19.246605 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="f884d097-5af4-48b5-8a84-173dcaeedf4e" containerName="collect-profiles" Dec 05 11:46:19.250529 master-0 kubenswrapper[24928]: I1205 11:46:19.246942 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="f884d097-5af4-48b5-8a84-173dcaeedf4e" containerName="collect-profiles" Dec 05 11:46:19.250529 master-0 kubenswrapper[24928]: I1205 11:46:19.249106 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:19.268650 master-0 kubenswrapper[24928]: I1205 11:46:19.268584 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-crjkl"] Dec 05 11:46:19.328125 master-0 kubenswrapper[24928]: I1205 11:46:19.328047 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs66t\" (UniqueName: \"kubernetes.io/projected/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-kube-api-access-zs66t\") pod \"certified-operators-crjkl\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:19.328370 master-0 kubenswrapper[24928]: I1205 11:46:19.328150 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-utilities\") pod \"certified-operators-crjkl\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:19.328370 master-0 kubenswrapper[24928]: I1205 11:46:19.328179 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-catalog-content\") pod \"certified-operators-crjkl\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:19.429837 master-0 kubenswrapper[24928]: I1205 11:46:19.429755 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs66t\" (UniqueName: \"kubernetes.io/projected/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-kube-api-access-zs66t\") pod \"certified-operators-crjkl\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:19.430086 master-0 kubenswrapper[24928]: I1205 11:46:19.429874 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-utilities\") pod \"certified-operators-crjkl\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:19.430086 master-0 kubenswrapper[24928]: I1205 11:46:19.429911 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-catalog-content\") pod \"certified-operators-crjkl\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:19.430798 master-0 kubenswrapper[24928]: I1205 11:46:19.430755 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-catalog-content\") pod \"certified-operators-crjkl\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:19.430798 master-0 kubenswrapper[24928]: I1205 11:46:19.430780 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-utilities\") pod \"certified-operators-crjkl\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:19.449297 master-0 kubenswrapper[24928]: I1205 11:46:19.449190 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs66t\" (UniqueName: \"kubernetes.io/projected/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-kube-api-access-zs66t\") pod \"certified-operators-crjkl\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:19.584374 master-0 kubenswrapper[24928]: I1205 11:46:19.584313 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:20.196950 master-0 kubenswrapper[24928]: I1205 11:46:20.196861 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-crjkl"] Dec 05 11:46:21.024410 master-0 kubenswrapper[24928]: I1205 11:46:21.024249 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crjkl" event={"ID":"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185","Type":"ContainerStarted","Data":"0583bac0995f062a0652569886ea9461d6055666653873286f82307c879e80bf"} Dec 05 11:46:22.039231 master-0 kubenswrapper[24928]: I1205 11:46:22.039140 24928 generic.go:334] "Generic (PLEG): container finished" podID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" containerID="26c6ac8c43b8b859f42dead40a2cd2478200743f044b15e5347f4c00a9310b14" exitCode=0 Dec 05 11:46:22.039231 master-0 kubenswrapper[24928]: I1205 11:46:22.039217 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crjkl" event={"ID":"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185","Type":"ContainerDied","Data":"26c6ac8c43b8b859f42dead40a2cd2478200743f044b15e5347f4c00a9310b14"} Dec 05 11:46:23.060196 master-0 kubenswrapper[24928]: I1205 11:46:23.059988 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crjkl" event={"ID":"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185","Type":"ContainerStarted","Data":"45436e95ea7e8329e1478bb66389aeee868a404228e9ea6e9c5e56eee2dec0a7"} Dec 05 11:46:24.080389 master-0 kubenswrapper[24928]: I1205 11:46:24.080275 24928 generic.go:334] "Generic (PLEG): container finished" podID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" containerID="45436e95ea7e8329e1478bb66389aeee868a404228e9ea6e9c5e56eee2dec0a7" exitCode=0 Dec 05 11:46:24.080389 master-0 kubenswrapper[24928]: I1205 11:46:24.080339 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crjkl" event={"ID":"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185","Type":"ContainerDied","Data":"45436e95ea7e8329e1478bb66389aeee868a404228e9ea6e9c5e56eee2dec0a7"} Dec 05 11:46:26.107874 master-0 kubenswrapper[24928]: I1205 11:46:26.107809 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crjkl" event={"ID":"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185","Type":"ContainerStarted","Data":"129e70f9c60d88e3f52b50f27e559601fabc893fa3c99f60008c630a6e9761b5"} Dec 05 11:46:26.139510 master-0 kubenswrapper[24928]: I1205 11:46:26.139356 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-crjkl" podStartSLOduration=4.301456525 podStartE2EDuration="7.139331709s" podCreationTimestamp="2025-12-05 11:46:19 +0000 UTC" firstStartedPulling="2025-12-05 11:46:22.043265645 +0000 UTC m=+3542.046459526" lastFinishedPulling="2025-12-05 11:46:24.881140859 +0000 UTC m=+3544.884334710" observedRunningTime="2025-12-05 11:46:26.127756627 +0000 UTC m=+3546.130950518" watchObservedRunningTime="2025-12-05 11:46:26.139331709 +0000 UTC m=+3546.142525560" Dec 05 11:46:29.585323 master-0 kubenswrapper[24928]: I1205 11:46:29.585224 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:29.585323 master-0 kubenswrapper[24928]: I1205 11:46:29.585333 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:29.639837 master-0 kubenswrapper[24928]: I1205 11:46:29.639774 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:30.251384 master-0 kubenswrapper[24928]: I1205 11:46:30.251332 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:30.998231 master-0 kubenswrapper[24928]: I1205 11:46:30.998133 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-crjkl"] Dec 05 11:46:32.225500 master-0 kubenswrapper[24928]: I1205 11:46:32.225389 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-crjkl" podUID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" containerName="registry-server" containerID="cri-o://129e70f9c60d88e3f52b50f27e559601fabc893fa3c99f60008c630a6e9761b5" gracePeriod=2 Dec 05 11:46:33.238842 master-0 kubenswrapper[24928]: I1205 11:46:33.238767 24928 generic.go:334] "Generic (PLEG): container finished" podID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" containerID="129e70f9c60d88e3f52b50f27e559601fabc893fa3c99f60008c630a6e9761b5" exitCode=0 Dec 05 11:46:33.238842 master-0 kubenswrapper[24928]: I1205 11:46:33.238827 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crjkl" event={"ID":"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185","Type":"ContainerDied","Data":"129e70f9c60d88e3f52b50f27e559601fabc893fa3c99f60008c630a6e9761b5"} Dec 05 11:46:33.239412 master-0 kubenswrapper[24928]: I1205 11:46:33.238871 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-crjkl" event={"ID":"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185","Type":"ContainerDied","Data":"0583bac0995f062a0652569886ea9461d6055666653873286f82307c879e80bf"} Dec 05 11:46:33.239412 master-0 kubenswrapper[24928]: I1205 11:46:33.238910 24928 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0583bac0995f062a0652569886ea9461d6055666653873286f82307c879e80bf" Dec 05 11:46:33.289342 master-0 kubenswrapper[24928]: I1205 11:46:33.289290 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:33.318486 master-0 kubenswrapper[24928]: I1205 11:46:33.318406 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-catalog-content\") pod \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " Dec 05 11:46:33.318714 master-0 kubenswrapper[24928]: I1205 11:46:33.318594 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zs66t\" (UniqueName: \"kubernetes.io/projected/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-kube-api-access-zs66t\") pod \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " Dec 05 11:46:33.318766 master-0 kubenswrapper[24928]: I1205 11:46:33.318733 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-utilities\") pod \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\" (UID: \"418646c9-0fd5-4b1c-ac1f-d45ec8ae5185\") " Dec 05 11:46:33.319775 master-0 kubenswrapper[24928]: I1205 11:46:33.319725 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-utilities" (OuterVolumeSpecName: "utilities") pod "418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" (UID: "418646c9-0fd5-4b1c-ac1f-d45ec8ae5185"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:46:33.320513 master-0 kubenswrapper[24928]: I1205 11:46:33.320401 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:46:33.368278 master-0 kubenswrapper[24928]: I1205 11:46:33.368177 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-kube-api-access-zs66t" (OuterVolumeSpecName: "kube-api-access-zs66t") pod "418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" (UID: "418646c9-0fd5-4b1c-ac1f-d45ec8ae5185"). InnerVolumeSpecName "kube-api-access-zs66t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:46:33.372271 master-0 kubenswrapper[24928]: I1205 11:46:33.372244 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" (UID: "418646c9-0fd5-4b1c-ac1f-d45ec8ae5185"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:46:33.422281 master-0 kubenswrapper[24928]: I1205 11:46:33.422211 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zs66t\" (UniqueName: \"kubernetes.io/projected/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-kube-api-access-zs66t\") on node \"master-0\" DevicePath \"\"" Dec 05 11:46:33.422281 master-0 kubenswrapper[24928]: I1205 11:46:33.422270 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:46:34.252405 master-0 kubenswrapper[24928]: I1205 11:46:34.252320 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-crjkl" Dec 05 11:46:34.300164 master-0 kubenswrapper[24928]: I1205 11:46:34.300080 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-crjkl"] Dec 05 11:46:34.317835 master-0 kubenswrapper[24928]: I1205 11:46:34.317732 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-crjkl"] Dec 05 11:46:36.216904 master-0 kubenswrapper[24928]: I1205 11:46:36.216835 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" path="/var/lib/kubelet/pods/418646c9-0fd5-4b1c-ac1f-d45ec8ae5185/volumes" Dec 05 11:51:06.933230 master-0 kubenswrapper[24928]: I1205 11:51:06.933140 24928 trace.go:236] Trace[1969192579]: "Calculate volume metrics of mysql-db for pod openstack/openstack-cell1-galera-0" (05-Dec-2025 11:51:05.619) (total time: 1313ms): Dec 05 11:51:06.933230 master-0 kubenswrapper[24928]: Trace[1969192579]: [1.313277218s] [1.313277218s] END Dec 05 11:52:49.993707 master-0 kubenswrapper[24928]: I1205 11:52:49.993560 24928 scope.go:117] "RemoveContainer" containerID="45436e95ea7e8329e1478bb66389aeee868a404228e9ea6e9c5e56eee2dec0a7" Dec 05 11:52:50.033647 master-0 kubenswrapper[24928]: I1205 11:52:50.033595 24928 scope.go:117] "RemoveContainer" containerID="129e70f9c60d88e3f52b50f27e559601fabc893fa3c99f60008c630a6e9761b5" Dec 05 11:52:50.073757 master-0 kubenswrapper[24928]: I1205 11:52:50.073701 24928 scope.go:117] "RemoveContainer" containerID="26c6ac8c43b8b859f42dead40a2cd2478200743f044b15e5347f4c00a9310b14" Dec 05 11:53:12.632695 master-0 kubenswrapper[24928]: I1205 11:53:12.632613 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h7tv2/must-gather-jbb7m"] Dec 05 11:53:12.633522 master-0 kubenswrapper[24928]: E1205 11:53:12.633296 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" containerName="extract-utilities" Dec 05 11:53:12.633522 master-0 kubenswrapper[24928]: I1205 11:53:12.633313 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" containerName="extract-utilities" Dec 05 11:53:12.633522 master-0 kubenswrapper[24928]: E1205 11:53:12.633327 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" containerName="extract-content" Dec 05 11:53:12.633522 master-0 kubenswrapper[24928]: I1205 11:53:12.633334 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" containerName="extract-content" Dec 05 11:53:12.633522 master-0 kubenswrapper[24928]: E1205 11:53:12.633359 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" containerName="registry-server" Dec 05 11:53:12.633522 master-0 kubenswrapper[24928]: I1205 11:53:12.633366 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" containerName="registry-server" Dec 05 11:53:12.633816 master-0 kubenswrapper[24928]: I1205 11:53:12.633625 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="418646c9-0fd5-4b1c-ac1f-d45ec8ae5185" containerName="registry-server" Dec 05 11:53:12.635188 master-0 kubenswrapper[24928]: I1205 11:53:12.635145 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/must-gather-jbb7m" Dec 05 11:53:12.640551 master-0 kubenswrapper[24928]: I1205 11:53:12.638958 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h7tv2"/"openshift-service-ca.crt" Dec 05 11:53:12.640551 master-0 kubenswrapper[24928]: I1205 11:53:12.639371 24928 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h7tv2"/"kube-root-ca.crt" Dec 05 11:53:12.652379 master-0 kubenswrapper[24928]: I1205 11:53:12.650549 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h7tv2/must-gather-t65nv"] Dec 05 11:53:12.652638 master-0 kubenswrapper[24928]: I1205 11:53:12.652604 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/must-gather-t65nv" Dec 05 11:53:12.665079 master-0 kubenswrapper[24928]: I1205 11:53:12.664696 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h7tv2/must-gather-jbb7m"] Dec 05 11:53:12.719507 master-0 kubenswrapper[24928]: I1205 11:53:12.719436 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h7tv2/must-gather-t65nv"] Dec 05 11:53:12.746630 master-0 kubenswrapper[24928]: I1205 11:53:12.742293 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kq75d\" (UniqueName: \"kubernetes.io/projected/8c4d353a-4199-4c65-8e5c-c8ad348bbb3d-kube-api-access-kq75d\") pod \"must-gather-jbb7m\" (UID: \"8c4d353a-4199-4c65-8e5c-c8ad348bbb3d\") " pod="openshift-must-gather-h7tv2/must-gather-jbb7m" Dec 05 11:53:12.746630 master-0 kubenswrapper[24928]: I1205 11:53:12.742380 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8c4d353a-4199-4c65-8e5c-c8ad348bbb3d-must-gather-output\") pod \"must-gather-jbb7m\" (UID: \"8c4d353a-4199-4c65-8e5c-c8ad348bbb3d\") " pod="openshift-must-gather-h7tv2/must-gather-jbb7m" Dec 05 11:53:12.844653 master-0 kubenswrapper[24928]: I1205 11:53:12.844565 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c8gd\" (UniqueName: \"kubernetes.io/projected/b35bd53c-939c-48bb-b46b-fa842ff0859c-kube-api-access-8c8gd\") pod \"must-gather-t65nv\" (UID: \"b35bd53c-939c-48bb-b46b-fa842ff0859c\") " pod="openshift-must-gather-h7tv2/must-gather-t65nv" Dec 05 11:53:12.844885 master-0 kubenswrapper[24928]: I1205 11:53:12.844829 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kq75d\" (UniqueName: \"kubernetes.io/projected/8c4d353a-4199-4c65-8e5c-c8ad348bbb3d-kube-api-access-kq75d\") pod \"must-gather-jbb7m\" (UID: \"8c4d353a-4199-4c65-8e5c-c8ad348bbb3d\") " pod="openshift-must-gather-h7tv2/must-gather-jbb7m" Dec 05 11:53:12.844885 master-0 kubenswrapper[24928]: I1205 11:53:12.844874 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8c4d353a-4199-4c65-8e5c-c8ad348bbb3d-must-gather-output\") pod \"must-gather-jbb7m\" (UID: \"8c4d353a-4199-4c65-8e5c-c8ad348bbb3d\") " pod="openshift-must-gather-h7tv2/must-gather-jbb7m" Dec 05 11:53:12.844976 master-0 kubenswrapper[24928]: I1205 11:53:12.844952 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b35bd53c-939c-48bb-b46b-fa842ff0859c-must-gather-output\") pod \"must-gather-t65nv\" (UID: \"b35bd53c-939c-48bb-b46b-fa842ff0859c\") " pod="openshift-must-gather-h7tv2/must-gather-t65nv" Dec 05 11:53:12.845896 master-0 kubenswrapper[24928]: I1205 11:53:12.845854 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/8c4d353a-4199-4c65-8e5c-c8ad348bbb3d-must-gather-output\") pod \"must-gather-jbb7m\" (UID: \"8c4d353a-4199-4c65-8e5c-c8ad348bbb3d\") " pod="openshift-must-gather-h7tv2/must-gather-jbb7m" Dec 05 11:53:12.860819 master-0 kubenswrapper[24928]: I1205 11:53:12.860720 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kq75d\" (UniqueName: \"kubernetes.io/projected/8c4d353a-4199-4c65-8e5c-c8ad348bbb3d-kube-api-access-kq75d\") pod \"must-gather-jbb7m\" (UID: \"8c4d353a-4199-4c65-8e5c-c8ad348bbb3d\") " pod="openshift-must-gather-h7tv2/must-gather-jbb7m" Dec 05 11:53:12.947181 master-0 kubenswrapper[24928]: I1205 11:53:12.947031 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b35bd53c-939c-48bb-b46b-fa842ff0859c-must-gather-output\") pod \"must-gather-t65nv\" (UID: \"b35bd53c-939c-48bb-b46b-fa842ff0859c\") " pod="openshift-must-gather-h7tv2/must-gather-t65nv" Dec 05 11:53:12.947181 master-0 kubenswrapper[24928]: I1205 11:53:12.947141 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c8gd\" (UniqueName: \"kubernetes.io/projected/b35bd53c-939c-48bb-b46b-fa842ff0859c-kube-api-access-8c8gd\") pod \"must-gather-t65nv\" (UID: \"b35bd53c-939c-48bb-b46b-fa842ff0859c\") " pod="openshift-must-gather-h7tv2/must-gather-t65nv" Dec 05 11:53:12.947797 master-0 kubenswrapper[24928]: I1205 11:53:12.947721 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/b35bd53c-939c-48bb-b46b-fa842ff0859c-must-gather-output\") pod \"must-gather-t65nv\" (UID: \"b35bd53c-939c-48bb-b46b-fa842ff0859c\") " pod="openshift-must-gather-h7tv2/must-gather-t65nv" Dec 05 11:53:12.965309 master-0 kubenswrapper[24928]: I1205 11:53:12.964869 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c8gd\" (UniqueName: \"kubernetes.io/projected/b35bd53c-939c-48bb-b46b-fa842ff0859c-kube-api-access-8c8gd\") pod \"must-gather-t65nv\" (UID: \"b35bd53c-939c-48bb-b46b-fa842ff0859c\") " pod="openshift-must-gather-h7tv2/must-gather-t65nv" Dec 05 11:53:12.971211 master-0 kubenswrapper[24928]: I1205 11:53:12.968142 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/must-gather-jbb7m" Dec 05 11:53:12.978679 master-0 kubenswrapper[24928]: I1205 11:53:12.978594 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/must-gather-t65nv" Dec 05 11:53:13.210708 master-0 kubenswrapper[24928]: I1205 11:53:13.210295 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x8mtr"] Dec 05 11:53:13.216406 master-0 kubenswrapper[24928]: I1205 11:53:13.213610 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:13.222242 master-0 kubenswrapper[24928]: I1205 11:53:13.221745 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x8mtr"] Dec 05 11:53:13.275688 master-0 kubenswrapper[24928]: I1205 11:53:13.273929 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-catalog-content\") pod \"community-operators-x8mtr\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:13.275688 master-0 kubenswrapper[24928]: I1205 11:53:13.274041 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-utilities\") pod \"community-operators-x8mtr\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:13.275688 master-0 kubenswrapper[24928]: I1205 11:53:13.274221 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x4s8\" (UniqueName: \"kubernetes.io/projected/7e52571a-e455-49af-aba7-457d90e0a1fa-kube-api-access-8x4s8\") pod \"community-operators-x8mtr\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:13.390883 master-0 kubenswrapper[24928]: I1205 11:53:13.390430 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-catalog-content\") pod \"community-operators-x8mtr\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:13.390883 master-0 kubenswrapper[24928]: I1205 11:53:13.390486 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-utilities\") pod \"community-operators-x8mtr\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:13.390883 master-0 kubenswrapper[24928]: I1205 11:53:13.390553 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x4s8\" (UniqueName: \"kubernetes.io/projected/7e52571a-e455-49af-aba7-457d90e0a1fa-kube-api-access-8x4s8\") pod \"community-operators-x8mtr\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:13.391233 master-0 kubenswrapper[24928]: I1205 11:53:13.391030 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-utilities\") pod \"community-operators-x8mtr\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:13.391233 master-0 kubenswrapper[24928]: I1205 11:53:13.391039 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-catalog-content\") pod \"community-operators-x8mtr\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:13.422836 master-0 kubenswrapper[24928]: I1205 11:53:13.422706 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x4s8\" (UniqueName: \"kubernetes.io/projected/7e52571a-e455-49af-aba7-457d90e0a1fa-kube-api-access-8x4s8\") pod \"community-operators-x8mtr\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:13.551331 master-0 kubenswrapper[24928]: I1205 11:53:13.551080 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:13.608176 master-0 kubenswrapper[24928]: I1205 11:53:13.608097 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h7tv2/must-gather-jbb7m"] Dec 05 11:53:13.633485 master-0 kubenswrapper[24928]: I1205 11:53:13.633443 24928 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 11:53:13.713458 master-0 kubenswrapper[24928]: I1205 11:53:13.711638 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h7tv2/must-gather-t65nv"] Dec 05 11:53:14.165828 master-0 kubenswrapper[24928]: W1205 11:53:14.164747 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e52571a_e455_49af_aba7_457d90e0a1fa.slice/crio-7ed2afe4cd6dcaf09251589cf06944c1fc24ead0641c15ec16fbbb9efcd32914 WatchSource:0}: Error finding container 7ed2afe4cd6dcaf09251589cf06944c1fc24ead0641c15ec16fbbb9efcd32914: Status 404 returned error can't find the container with id 7ed2afe4cd6dcaf09251589cf06944c1fc24ead0641c15ec16fbbb9efcd32914 Dec 05 11:53:14.167739 master-0 kubenswrapper[24928]: I1205 11:53:14.166496 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x8mtr"] Dec 05 11:53:14.449610 master-0 kubenswrapper[24928]: I1205 11:53:14.449469 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h7tv2/must-gather-jbb7m" event={"ID":"8c4d353a-4199-4c65-8e5c-c8ad348bbb3d","Type":"ContainerStarted","Data":"8a8a9925a2b50d62f93077c764ecafc0f24aecd18622a72314fb421cec7a4de4"} Dec 05 11:53:14.451920 master-0 kubenswrapper[24928]: I1205 11:53:14.451882 24928 generic.go:334] "Generic (PLEG): container finished" podID="7e52571a-e455-49af-aba7-457d90e0a1fa" containerID="dd066b4cf40ae32f522f00478ad82c3743d5da64b11516a9c3d42fe97a8810aa" exitCode=0 Dec 05 11:53:14.452040 master-0 kubenswrapper[24928]: I1205 11:53:14.451984 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8mtr" event={"ID":"7e52571a-e455-49af-aba7-457d90e0a1fa","Type":"ContainerDied","Data":"dd066b4cf40ae32f522f00478ad82c3743d5da64b11516a9c3d42fe97a8810aa"} Dec 05 11:53:14.452040 master-0 kubenswrapper[24928]: I1205 11:53:14.452016 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8mtr" event={"ID":"7e52571a-e455-49af-aba7-457d90e0a1fa","Type":"ContainerStarted","Data":"7ed2afe4cd6dcaf09251589cf06944c1fc24ead0641c15ec16fbbb9efcd32914"} Dec 05 11:53:14.454911 master-0 kubenswrapper[24928]: I1205 11:53:14.454850 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h7tv2/must-gather-t65nv" event={"ID":"b35bd53c-939c-48bb-b46b-fa842ff0859c","Type":"ContainerStarted","Data":"1cc3d7db921101c65a301a478176e419c2d40129737eb6347aab3ba7d3e624b6"} Dec 05 11:53:15.469526 master-0 kubenswrapper[24928]: I1205 11:53:15.469386 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h7tv2/must-gather-t65nv" event={"ID":"b35bd53c-939c-48bb-b46b-fa842ff0859c","Type":"ContainerStarted","Data":"edc42afa97ab55ece9041dff49e4735dd5c67c9db75bffa56488f3ef874bc55a"} Dec 05 11:53:15.890793 master-0 kubenswrapper[24928]: I1205 11:53:15.890640 24928 ???:1] "http: TLS handshake error from 192.168.32.10:34570: EOF" Dec 05 11:53:16.045338 master-0 kubenswrapper[24928]: E1205 11:53:16.045253 24928 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e52571a_e455_49af_aba7_457d90e0a1fa.slice/crio-b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e52571a_e455_49af_aba7_457d90e0a1fa.slice/crio-conmon-b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891.scope\": RecentStats: unable to find data in memory cache]" Dec 05 11:53:16.491794 master-0 kubenswrapper[24928]: I1205 11:53:16.489407 24928 generic.go:334] "Generic (PLEG): container finished" podID="7e52571a-e455-49af-aba7-457d90e0a1fa" containerID="b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891" exitCode=0 Dec 05 11:53:16.491794 master-0 kubenswrapper[24928]: I1205 11:53:16.489520 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8mtr" event={"ID":"7e52571a-e455-49af-aba7-457d90e0a1fa","Type":"ContainerDied","Data":"b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891"} Dec 05 11:53:16.494303 master-0 kubenswrapper[24928]: I1205 11:53:16.494249 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h7tv2/must-gather-t65nv" event={"ID":"b35bd53c-939c-48bb-b46b-fa842ff0859c","Type":"ContainerStarted","Data":"6319b630f327c89dc123f56b5183a15f2024746de493435ddd63da279b468e01"} Dec 05 11:53:16.532845 master-0 kubenswrapper[24928]: I1205 11:53:16.532772 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h7tv2/must-gather-t65nv" podStartSLOduration=3.197963526 podStartE2EDuration="4.532756069s" podCreationTimestamp="2025-12-05 11:53:12 +0000 UTC" firstStartedPulling="2025-12-05 11:53:13.703857646 +0000 UTC m=+3953.707051497" lastFinishedPulling="2025-12-05 11:53:15.038650189 +0000 UTC m=+3955.041844040" observedRunningTime="2025-12-05 11:53:16.532701648 +0000 UTC m=+3956.535895509" watchObservedRunningTime="2025-12-05 11:53:16.532756069 +0000 UTC m=+3956.535949920" Dec 05 11:53:18.475915 master-0 kubenswrapper[24928]: I1205 11:53:18.475296 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-version_cluster-version-operator-6d5d5dcc89-27xm6_1fd0a349-0f66-4c85-95e4-8bbc96648c05/cluster-version-operator/0.log" Dec 05 11:53:18.524739 master-0 kubenswrapper[24928]: I1205 11:53:18.524681 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h7tv2/must-gather-jbb7m" event={"ID":"8c4d353a-4199-4c65-8e5c-c8ad348bbb3d","Type":"ContainerStarted","Data":"583c3982876fbe04d0bb77ef6ba47692caf59055b41377169d6dd0e9b93d09b2"} Dec 05 11:53:19.572127 master-0 kubenswrapper[24928]: I1205 11:53:19.572073 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h7tv2/must-gather-jbb7m" event={"ID":"8c4d353a-4199-4c65-8e5c-c8ad348bbb3d","Type":"ContainerStarted","Data":"31dc5c0ce0d13575fb1c90b69d175744c08c4a5edf456bb6839f7fa116ea1f58"} Dec 05 11:53:19.655841 master-0 kubenswrapper[24928]: I1205 11:53:19.601274 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h7tv2/must-gather-jbb7m" podStartSLOduration=3.21445961 podStartE2EDuration="7.601255493s" podCreationTimestamp="2025-12-05 11:53:12 +0000 UTC" firstStartedPulling="2025-12-05 11:53:13.633351072 +0000 UTC m=+3953.636544923" lastFinishedPulling="2025-12-05 11:53:18.020146955 +0000 UTC m=+3958.023340806" observedRunningTime="2025-12-05 11:53:19.600037263 +0000 UTC m=+3959.603231114" watchObservedRunningTime="2025-12-05 11:53:19.601255493 +0000 UTC m=+3959.604449344" Dec 05 11:53:19.676639 master-0 kubenswrapper[24928]: I1205 11:53:19.676571 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8mtr" event={"ID":"7e52571a-e455-49af-aba7-457d90e0a1fa","Type":"ContainerStarted","Data":"9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b"} Dec 05 11:53:19.713247 master-0 kubenswrapper[24928]: I1205 11:53:19.712563 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x8mtr" podStartSLOduration=2.8130090770000002 podStartE2EDuration="6.712539444s" podCreationTimestamp="2025-12-05 11:53:13 +0000 UTC" firstStartedPulling="2025-12-05 11:53:14.454151495 +0000 UTC m=+3954.457345346" lastFinishedPulling="2025-12-05 11:53:18.353681862 +0000 UTC m=+3958.356875713" observedRunningTime="2025-12-05 11:53:19.704759584 +0000 UTC m=+3959.707953435" watchObservedRunningTime="2025-12-05 11:53:19.712539444 +0000 UTC m=+3959.715733295" Dec 05 11:53:22.923320 master-0 kubenswrapper[24928]: I1205 11:53:22.923129 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-fpl59_2a019084-2c01-4311-81d4-df4e5fe47107/controller/0.log" Dec 05 11:53:23.553915 master-0 kubenswrapper[24928]: I1205 11:53:23.553812 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:23.553915 master-0 kubenswrapper[24928]: I1205 11:53:23.553912 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:23.609403 master-0 kubenswrapper[24928]: I1205 11:53:23.609325 24928 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:23.802127 master-0 kubenswrapper[24928]: I1205 11:53:23.802041 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:24.010852 master-0 kubenswrapper[24928]: I1205 11:53:24.010220 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-hvbdn_381f465d-c34b-46c0-99f6-d50a6f6307bb/nmstate-console-plugin/0.log" Dec 05 11:53:24.010852 master-0 kubenswrapper[24928]: I1205 11:53:24.010301 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-fpl59_2a019084-2c01-4311-81d4-df4e5fe47107/kube-rbac-proxy/0.log" Dec 05 11:53:24.086219 master-0 kubenswrapper[24928]: I1205 11:53:24.083214 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-hxkln_27c2f18f-89e8-4439-a383-6bc8f6bed3af/nmstate-handler/0.log" Dec 05 11:53:24.137041 master-0 kubenswrapper[24928]: I1205 11:53:24.136931 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/controller/0.log" Dec 05 11:53:24.148445 master-0 kubenswrapper[24928]: I1205 11:53:24.145212 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-jqhpk_1501ec1c-63ed-41c9-8751-affef02bd884/nmstate-metrics/0.log" Dec 05 11:53:24.176207 master-0 kubenswrapper[24928]: I1205 11:53:24.175724 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-jqhpk_1501ec1c-63ed-41c9-8751-affef02bd884/kube-rbac-proxy/0.log" Dec 05 11:53:24.268524 master-0 kubenswrapper[24928]: I1205 11:53:24.264706 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-5fcz7_c79e22f5-57e7-437b-bdb4-52cc79cd73b0/nmstate-operator/0.log" Dec 05 11:53:24.345469 master-0 kubenswrapper[24928]: I1205 11:53:24.343575 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-jxlrb_4c719724-d364-4e6b-8c86-b72d3152536d/nmstate-webhook/0.log" Dec 05 11:53:24.371466 master-0 kubenswrapper[24928]: I1205 11:53:24.371335 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x8mtr"] Dec 05 11:53:25.720941 master-0 kubenswrapper[24928]: I1205 11:53:25.720777 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/frr/0.log" Dec 05 11:53:25.730902 master-0 kubenswrapper[24928]: I1205 11:53:25.730692 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/reloader/0.log" Dec 05 11:53:25.737439 master-0 kubenswrapper[24928]: I1205 11:53:25.737367 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/frr-metrics/0.log" Dec 05 11:53:25.746803 master-0 kubenswrapper[24928]: I1205 11:53:25.746749 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/kube-rbac-proxy/0.log" Dec 05 11:53:25.758752 master-0 kubenswrapper[24928]: I1205 11:53:25.758667 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/kube-rbac-proxy-frr/0.log" Dec 05 11:53:25.766269 master-0 kubenswrapper[24928]: I1205 11:53:25.766213 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/cp-frr-files/0.log" Dec 05 11:53:25.777596 master-0 kubenswrapper[24928]: I1205 11:53:25.777503 24928 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x8mtr" podUID="7e52571a-e455-49af-aba7-457d90e0a1fa" containerName="registry-server" containerID="cri-o://9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b" gracePeriod=2 Dec 05 11:53:25.788073 master-0 kubenswrapper[24928]: I1205 11:53:25.788003 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/cp-reloader/0.log" Dec 05 11:53:25.795235 master-0 kubenswrapper[24928]: I1205 11:53:25.795180 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/cp-metrics/0.log" Dec 05 11:53:25.808917 master-0 kubenswrapper[24928]: I1205 11:53:25.808859 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-dlsnb_a65480f6-aa9d-45b5-b3ba-541e87b7b145/frr-k8s-webhook-server/0.log" Dec 05 11:53:25.830687 master-0 kubenswrapper[24928]: I1205 11:53:25.830622 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9d5bd9bc7-q878m_697f986f-bfa7-4e85-9d4b-5f07a4675228/manager/0.log" Dec 05 11:53:25.842880 master-0 kubenswrapper[24928]: I1205 11:53:25.842824 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5f77dd7bb4-xmg4x_c39fe00d-1550-4d88-8b8a-95439d6e58e1/webhook-server/0.log" Dec 05 11:53:26.165607 master-0 kubenswrapper[24928]: I1205 11:53:26.165530 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h7tv2/master-0-debug-629dn"] Dec 05 11:53:26.217062 master-0 kubenswrapper[24928]: I1205 11:53:26.217004 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/master-0-debug-629dn" Dec 05 11:53:26.256732 master-0 kubenswrapper[24928]: I1205 11:53:26.256639 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d44374e-1ad4-47df-a596-e8c2b974783a-host\") pod \"master-0-debug-629dn\" (UID: \"6d44374e-1ad4-47df-a596-e8c2b974783a\") " pod="openshift-must-gather-h7tv2/master-0-debug-629dn" Dec 05 11:53:26.257286 master-0 kubenswrapper[24928]: I1205 11:53:26.257229 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mntvb\" (UniqueName: \"kubernetes.io/projected/6d44374e-1ad4-47df-a596-e8c2b974783a-kube-api-access-mntvb\") pod \"master-0-debug-629dn\" (UID: \"6d44374e-1ad4-47df-a596-e8c2b974783a\") " pod="openshift-must-gather-h7tv2/master-0-debug-629dn" Dec 05 11:53:26.359396 master-0 kubenswrapper[24928]: I1205 11:53:26.359333 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d44374e-1ad4-47df-a596-e8c2b974783a-host\") pod \"master-0-debug-629dn\" (UID: \"6d44374e-1ad4-47df-a596-e8c2b974783a\") " pod="openshift-must-gather-h7tv2/master-0-debug-629dn" Dec 05 11:53:26.359577 master-0 kubenswrapper[24928]: I1205 11:53:26.359409 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mntvb\" (UniqueName: \"kubernetes.io/projected/6d44374e-1ad4-47df-a596-e8c2b974783a-kube-api-access-mntvb\") pod \"master-0-debug-629dn\" (UID: \"6d44374e-1ad4-47df-a596-e8c2b974783a\") " pod="openshift-must-gather-h7tv2/master-0-debug-629dn" Dec 05 11:53:26.359652 master-0 kubenswrapper[24928]: I1205 11:53:26.359603 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d44374e-1ad4-47df-a596-e8c2b974783a-host\") pod \"master-0-debug-629dn\" (UID: \"6d44374e-1ad4-47df-a596-e8c2b974783a\") " pod="openshift-must-gather-h7tv2/master-0-debug-629dn" Dec 05 11:53:26.381472 master-0 kubenswrapper[24928]: I1205 11:53:26.379718 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mntvb\" (UniqueName: \"kubernetes.io/projected/6d44374e-1ad4-47df-a596-e8c2b974783a-kube-api-access-mntvb\") pod \"master-0-debug-629dn\" (UID: \"6d44374e-1ad4-47df-a596-e8c2b974783a\") " pod="openshift-must-gather-h7tv2/master-0-debug-629dn" Dec 05 11:53:26.453275 master-0 kubenswrapper[24928]: I1205 11:53:26.453145 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9stls_b953bb78-052f-4b26-94d3-da9bab7a9fdc/speaker/0.log" Dec 05 11:53:26.466451 master-0 kubenswrapper[24928]: I1205 11:53:26.465142 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9stls_b953bb78-052f-4b26-94d3-da9bab7a9fdc/kube-rbac-proxy/0.log" Dec 05 11:53:26.480409 master-0 kubenswrapper[24928]: I1205 11:53:26.472991 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:26.565446 master-0 kubenswrapper[24928]: I1205 11:53:26.565246 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x4s8\" (UniqueName: \"kubernetes.io/projected/7e52571a-e455-49af-aba7-457d90e0a1fa-kube-api-access-8x4s8\") pod \"7e52571a-e455-49af-aba7-457d90e0a1fa\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " Dec 05 11:53:26.565673 master-0 kubenswrapper[24928]: I1205 11:53:26.565522 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-utilities\") pod \"7e52571a-e455-49af-aba7-457d90e0a1fa\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " Dec 05 11:53:26.565673 master-0 kubenswrapper[24928]: I1205 11:53:26.565577 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-catalog-content\") pod \"7e52571a-e455-49af-aba7-457d90e0a1fa\" (UID: \"7e52571a-e455-49af-aba7-457d90e0a1fa\") " Dec 05 11:53:26.570585 master-0 kubenswrapper[24928]: I1205 11:53:26.566840 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-utilities" (OuterVolumeSpecName: "utilities") pod "7e52571a-e455-49af-aba7-457d90e0a1fa" (UID: "7e52571a-e455-49af-aba7-457d90e0a1fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:53:26.577505 master-0 kubenswrapper[24928]: I1205 11:53:26.576753 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e52571a-e455-49af-aba7-457d90e0a1fa-kube-api-access-8x4s8" (OuterVolumeSpecName: "kube-api-access-8x4s8") pod "7e52571a-e455-49af-aba7-457d90e0a1fa" (UID: "7e52571a-e455-49af-aba7-457d90e0a1fa"). InnerVolumeSpecName "kube-api-access-8x4s8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:53:26.644859 master-0 kubenswrapper[24928]: I1205 11:53:26.643036 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/master-0-debug-629dn" Dec 05 11:53:26.681037 master-0 kubenswrapper[24928]: I1205 11:53:26.680978 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e52571a-e455-49af-aba7-457d90e0a1fa" (UID: "7e52571a-e455-49af-aba7-457d90e0a1fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 11:53:26.707690 master-0 kubenswrapper[24928]: I1205 11:53:26.707277 24928 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-utilities\") on node \"master-0\" DevicePath \"\"" Dec 05 11:53:26.707690 master-0 kubenswrapper[24928]: I1205 11:53:26.707315 24928 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e52571a-e455-49af-aba7-457d90e0a1fa-catalog-content\") on node \"master-0\" DevicePath \"\"" Dec 05 11:53:26.707690 master-0 kubenswrapper[24928]: I1205 11:53:26.707327 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x4s8\" (UniqueName: \"kubernetes.io/projected/7e52571a-e455-49af-aba7-457d90e0a1fa-kube-api-access-8x4s8\") on node \"master-0\" DevicePath \"\"" Dec 05 11:53:26.805945 master-0 kubenswrapper[24928]: I1205 11:53:26.805880 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h7tv2/master-0-debug-629dn" event={"ID":"6d44374e-1ad4-47df-a596-e8c2b974783a","Type":"ContainerStarted","Data":"0c094ed9554bf0de9b4511b995df9c8015ae6cc7ea4f9236d73d4d59a3260adc"} Dec 05 11:53:26.816465 master-0 kubenswrapper[24928]: I1205 11:53:26.814616 24928 generic.go:334] "Generic (PLEG): container finished" podID="7e52571a-e455-49af-aba7-457d90e0a1fa" containerID="9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b" exitCode=0 Dec 05 11:53:26.816465 master-0 kubenswrapper[24928]: I1205 11:53:26.814675 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8mtr" event={"ID":"7e52571a-e455-49af-aba7-457d90e0a1fa","Type":"ContainerDied","Data":"9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b"} Dec 05 11:53:26.816465 master-0 kubenswrapper[24928]: I1205 11:53:26.814704 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8mtr" event={"ID":"7e52571a-e455-49af-aba7-457d90e0a1fa","Type":"ContainerDied","Data":"7ed2afe4cd6dcaf09251589cf06944c1fc24ead0641c15ec16fbbb9efcd32914"} Dec 05 11:53:26.816465 master-0 kubenswrapper[24928]: I1205 11:53:26.814722 24928 scope.go:117] "RemoveContainer" containerID="9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b" Dec 05 11:53:26.816465 master-0 kubenswrapper[24928]: I1205 11:53:26.814887 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x8mtr" Dec 05 11:53:26.858212 master-0 kubenswrapper[24928]: I1205 11:53:26.858167 24928 scope.go:117] "RemoveContainer" containerID="b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891" Dec 05 11:53:26.908453 master-0 kubenswrapper[24928]: I1205 11:53:26.895496 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x8mtr"] Dec 05 11:53:26.921242 master-0 kubenswrapper[24928]: I1205 11:53:26.920655 24928 scope.go:117] "RemoveContainer" containerID="dd066b4cf40ae32f522f00478ad82c3743d5da64b11516a9c3d42fe97a8810aa" Dec 05 11:53:26.921242 master-0 kubenswrapper[24928]: I1205 11:53:26.920791 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x8mtr"] Dec 05 11:53:26.996658 master-0 kubenswrapper[24928]: I1205 11:53:26.995626 24928 scope.go:117] "RemoveContainer" containerID="9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b" Dec 05 11:53:26.996658 master-0 kubenswrapper[24928]: E1205 11:53:26.996311 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b\": container with ID starting with 9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b not found: ID does not exist" containerID="9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b" Dec 05 11:53:26.996658 master-0 kubenswrapper[24928]: I1205 11:53:26.996348 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b"} err="failed to get container status \"9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b\": rpc error: code = NotFound desc = could not find container \"9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b\": container with ID starting with 9c81e2f79895d1cc1863dba82dc8ac036df99b26d1dbdbca925851b005da0e0b not found: ID does not exist" Dec 05 11:53:26.996658 master-0 kubenswrapper[24928]: I1205 11:53:26.996370 24928 scope.go:117] "RemoveContainer" containerID="b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891" Dec 05 11:53:26.996828 master-0 kubenswrapper[24928]: E1205 11:53:26.996693 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891\": container with ID starting with b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891 not found: ID does not exist" containerID="b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891" Dec 05 11:53:26.996828 master-0 kubenswrapper[24928]: I1205 11:53:26.996711 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891"} err="failed to get container status \"b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891\": rpc error: code = NotFound desc = could not find container \"b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891\": container with ID starting with b998fe362a70c139ade77a1c0ddc47b73ad5d7feb8953b0ff3c51412e8822891 not found: ID does not exist" Dec 05 11:53:26.996828 master-0 kubenswrapper[24928]: I1205 11:53:26.996724 24928 scope.go:117] "RemoveContainer" containerID="dd066b4cf40ae32f522f00478ad82c3743d5da64b11516a9c3d42fe97a8810aa" Dec 05 11:53:27.000448 master-0 kubenswrapper[24928]: E1205 11:53:26.996917 24928 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd066b4cf40ae32f522f00478ad82c3743d5da64b11516a9c3d42fe97a8810aa\": container with ID starting with dd066b4cf40ae32f522f00478ad82c3743d5da64b11516a9c3d42fe97a8810aa not found: ID does not exist" containerID="dd066b4cf40ae32f522f00478ad82c3743d5da64b11516a9c3d42fe97a8810aa" Dec 05 11:53:27.000448 master-0 kubenswrapper[24928]: I1205 11:53:26.996948 24928 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd066b4cf40ae32f522f00478ad82c3743d5da64b11516a9c3d42fe97a8810aa"} err="failed to get container status \"dd066b4cf40ae32f522f00478ad82c3743d5da64b11516a9c3d42fe97a8810aa\": rpc error: code = NotFound desc = could not find container \"dd066b4cf40ae32f522f00478ad82c3743d5da64b11516a9c3d42fe97a8810aa\": container with ID starting with dd066b4cf40ae32f522f00478ad82c3743d5da64b11516a9c3d42fe97a8810aa not found: ID does not exist" Dec 05 11:53:27.752204 master-0 kubenswrapper[24928]: I1205 11:53:27.751593 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcdctl/0.log" Dec 05 11:53:28.222593 master-0 kubenswrapper[24928]: I1205 11:53:28.222313 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e52571a-e455-49af-aba7-457d90e0a1fa" path="/var/lib/kubelet/pods/7e52571a-e455-49af-aba7-457d90e0a1fa/volumes" Dec 05 11:53:28.390910 master-0 kubenswrapper[24928]: I1205 11:53:28.390799 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-775fbfd4bb-cxrjv_e61eff80-d16c-4636-837d-09237981ca7f/oauth-openshift/0.log" Dec 05 11:53:28.881889 master-0 kubenswrapper[24928]: I1205 11:53:28.881824 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd/0.log" Dec 05 11:53:28.905333 master-0 kubenswrapper[24928]: I1205 11:53:28.903824 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd-metrics/0.log" Dec 05 11:53:28.917804 master-0 kubenswrapper[24928]: I1205 11:53:28.917758 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd-readyz/0.log" Dec 05 11:53:28.932116 master-0 kubenswrapper[24928]: I1205 11:53:28.932060 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd-rev/0.log" Dec 05 11:53:29.043263 master-0 kubenswrapper[24928]: I1205 11:53:29.043193 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/setup/0.log" Dec 05 11:53:29.058943 master-0 kubenswrapper[24928]: I1205 11:53:29.058873 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd-ensure-env-vars/0.log" Dec 05 11:53:29.082770 master-0 kubenswrapper[24928]: I1205 11:53:29.082360 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd-resources-copy/0.log" Dec 05 11:53:29.178613 master-0 kubenswrapper[24928]: I1205 11:53:29.177833 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-1-master-0_e149cbe7-1ec3-4a06-af61-3a8906b8e9ef/installer/0.log" Dec 05 11:53:29.203754 master-0 kubenswrapper[24928]: I1205 11:53:29.203698 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-2-master-0_68058d25-5888-495f-8571-453f0bf919d9/installer/0.log" Dec 05 11:53:29.285908 master-0 kubenswrapper[24928]: I1205 11:53:29.285800 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-2-retry-1-master-0_ee0f69ec-1862-4016-bf9d-7edf6554f473/installer/0.log" Dec 05 11:53:29.924502 master-0 kubenswrapper[24928]: I1205 11:53:29.922856 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8_d95a56ba-c940-4e3e-aed6-d8c04f1871b6/authentication-operator/0.log" Dec 05 11:53:29.961262 master-0 kubenswrapper[24928]: I1205 11:53:29.960778 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication-operator_authentication-operator-6c968fdfdf-t7sl8_d95a56ba-c940-4e3e-aed6-d8c04f1871b6/authentication-operator/1.log" Dec 05 11:53:30.727450 master-0 kubenswrapper[24928]: I1205 11:53:30.727311 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/assisted-installer_assisted-installer-controller-pd4q6_f1d74f88-1419-431e-80da-26db419f050e/assisted-installer-controller/0.log" Dec 05 11:53:31.019612 master-0 kubenswrapper[24928]: I1205 11:53:31.019509 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5465c8b4db-s4c2f_445d75af-d072-4fa0-91a7-f3fa579b9ca9/router/2.log" Dec 05 11:53:31.027259 master-0 kubenswrapper[24928]: I1205 11:53:31.027196 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5465c8b4db-s4c2f_445d75af-d072-4fa0-91a7-f3fa579b9ca9/router/1.log" Dec 05 11:53:31.707451 master-0 kubenswrapper[24928]: I1205 11:53:31.701899 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc"] Dec 05 11:53:31.707451 master-0 kubenswrapper[24928]: E1205 11:53:31.703387 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e52571a-e455-49af-aba7-457d90e0a1fa" containerName="registry-server" Dec 05 11:53:31.707451 master-0 kubenswrapper[24928]: I1205 11:53:31.703412 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e52571a-e455-49af-aba7-457d90e0a1fa" containerName="registry-server" Dec 05 11:53:31.707451 master-0 kubenswrapper[24928]: E1205 11:53:31.703496 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e52571a-e455-49af-aba7-457d90e0a1fa" containerName="extract-utilities" Dec 05 11:53:31.707451 master-0 kubenswrapper[24928]: I1205 11:53:31.703509 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e52571a-e455-49af-aba7-457d90e0a1fa" containerName="extract-utilities" Dec 05 11:53:31.707451 master-0 kubenswrapper[24928]: E1205 11:53:31.703543 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e52571a-e455-49af-aba7-457d90e0a1fa" containerName="extract-content" Dec 05 11:53:31.707451 master-0 kubenswrapper[24928]: I1205 11:53:31.703553 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e52571a-e455-49af-aba7-457d90e0a1fa" containerName="extract-content" Dec 05 11:53:31.707451 master-0 kubenswrapper[24928]: I1205 11:53:31.704324 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e52571a-e455-49af-aba7-457d90e0a1fa" containerName="registry-server" Dec 05 11:53:31.730809 master-0 kubenswrapper[24928]: I1205 11:53:31.730625 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.740557 master-0 kubenswrapper[24928]: I1205 11:53:31.740493 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc"] Dec 05 11:53:31.864447 master-0 kubenswrapper[24928]: I1205 11:53:31.863662 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-lib-modules\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.864447 master-0 kubenswrapper[24928]: I1205 11:53:31.863721 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8bv4\" (UniqueName: \"kubernetes.io/projected/376ec169-3f8c-467a-845e-2fbe472e6497-kube-api-access-t8bv4\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.864447 master-0 kubenswrapper[24928]: I1205 11:53:31.863772 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-proc\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.864447 master-0 kubenswrapper[24928]: I1205 11:53:31.863789 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-podres\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.864447 master-0 kubenswrapper[24928]: I1205 11:53:31.864254 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-sys\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.967041 master-0 kubenswrapper[24928]: I1205 11:53:31.965824 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-lib-modules\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.967041 master-0 kubenswrapper[24928]: I1205 11:53:31.965884 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8bv4\" (UniqueName: \"kubernetes.io/projected/376ec169-3f8c-467a-845e-2fbe472e6497-kube-api-access-t8bv4\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.967041 master-0 kubenswrapper[24928]: I1205 11:53:31.965938 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-proc\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.967041 master-0 kubenswrapper[24928]: I1205 11:53:31.965958 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-podres\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.967041 master-0 kubenswrapper[24928]: I1205 11:53:31.966092 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-sys\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.967041 master-0 kubenswrapper[24928]: I1205 11:53:31.966234 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-sys\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.967041 master-0 kubenswrapper[24928]: I1205 11:53:31.966282 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proc\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-proc\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.967041 master-0 kubenswrapper[24928]: I1205 11:53:31.966326 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"podres\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-podres\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.967041 master-0 kubenswrapper[24928]: I1205 11:53:31.966377 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/376ec169-3f8c-467a-845e-2fbe472e6497-lib-modules\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:31.990124 master-0 kubenswrapper[24928]: I1205 11:53:31.990045 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8bv4\" (UniqueName: \"kubernetes.io/projected/376ec169-3f8c-467a-845e-2fbe472e6497-kube-api-access-t8bv4\") pod \"perf-node-gather-daemonset-tz9gc\" (UID: \"376ec169-3f8c-467a-845e-2fbe472e6497\") " pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:32.133560 master-0 kubenswrapper[24928]: I1205 11:53:32.132186 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:32.307873 master-0 kubenswrapper[24928]: I1205 11:53:32.307771 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-oauth-apiserver_apiserver-85b8f855df-8g52w_ca45c52e-fb30-4e7c-8c3f-e685c0909916/oauth-apiserver/0.log" Dec 05 11:53:32.327246 master-0 kubenswrapper[24928]: I1205 11:53:32.327118 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-oauth-apiserver_apiserver-85b8f855df-8g52w_ca45c52e-fb30-4e7c-8c3f-e685c0909916/fix-audit-permissions/0.log" Dec 05 11:53:32.668639 master-0 kubenswrapper[24928]: I1205 11:53:32.668482 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc"] Dec 05 11:53:32.930345 master-0 kubenswrapper[24928]: I1205 11:53:32.930203 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" event={"ID":"376ec169-3f8c-467a-845e-2fbe472e6497","Type":"ContainerStarted","Data":"43c5ab5fd65acfa8cf6a61c8262886e688e9290e39181036930dd9f301eb58ca"} Dec 05 11:53:32.996487 master-0 kubenswrapper[24928]: I1205 11:53:32.996393 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-autoscaler-operator-5f49d774cd-cfg5f_1f132702-304b-46ff-b428-6df1deeffec3/kube-rbac-proxy/0.log" Dec 05 11:53:33.073405 master-0 kubenswrapper[24928]: I1205 11:53:33.073353 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-autoscaler-operator-5f49d774cd-cfg5f_1f132702-304b-46ff-b428-6df1deeffec3/cluster-autoscaler-operator/0.log" Dec 05 11:53:33.104944 master-0 kubenswrapper[24928]: I1205 11:53:33.104899 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-baremetal-operator-78f758c7b9-6t2gm_48bd1d86-a6f2-439f-ab04-6a9a442bec42/cluster-baremetal-operator/2.log" Dec 05 11:53:33.106791 master-0 kubenswrapper[24928]: I1205 11:53:33.106763 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-baremetal-operator-78f758c7b9-6t2gm_48bd1d86-a6f2-439f-ab04-6a9a442bec42/cluster-baremetal-operator/1.log" Dec 05 11:53:33.126741 master-0 kubenswrapper[24928]: I1205 11:53:33.126675 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_cluster-baremetal-operator-78f758c7b9-6t2gm_48bd1d86-a6f2-439f-ab04-6a9a442bec42/baremetal-kube-rbac-proxy/0.log" Dec 05 11:53:33.148001 master-0 kubenswrapper[24928]: I1205 11:53:33.147940 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-7df95c79b5-qnq6t_24444474-aa9a-4a0f-8b4d-90f0009e0dc7/control-plane-machine-set-operator/0.log" Dec 05 11:53:33.148225 master-0 kubenswrapper[24928]: I1205 11:53:33.148031 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-7df95c79b5-qnq6t_24444474-aa9a-4a0f-8b4d-90f0009e0dc7/control-plane-machine-set-operator/1.log" Dec 05 11:53:33.171510 master-0 kubenswrapper[24928]: I1205 11:53:33.171359 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-88d48b57d-x7jfs_82464e92-4de2-48f0-8772-a489abb16898/kube-rbac-proxy/0.log" Dec 05 11:53:33.189136 master-0 kubenswrapper[24928]: I1205 11:53:33.189014 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-88d48b57d-x7jfs_82464e92-4de2-48f0-8772-a489abb16898/machine-api-operator/0.log" Dec 05 11:53:33.973061 master-0 kubenswrapper[24928]: I1205 11:53:33.972991 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" event={"ID":"376ec169-3f8c-467a-845e-2fbe472e6497","Type":"ContainerStarted","Data":"6af6c351bb9b891e07eeb4de8f2061de8c0e3004c55c274590c5ca91921121be"} Dec 05 11:53:33.974658 master-0 kubenswrapper[24928]: I1205 11:53:33.974634 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:34.012312 master-0 kubenswrapper[24928]: I1205 11:53:34.011666 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" podStartSLOduration=3.011646471 podStartE2EDuration="3.011646471s" podCreationTimestamp="2025-12-05 11:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 11:53:33.993751844 +0000 UTC m=+3973.996945705" watchObservedRunningTime="2025-12-05 11:53:34.011646471 +0000 UTC m=+3974.014840322" Dec 05 11:53:34.528775 master-0 kubenswrapper[24928]: I1205 11:53:34.528720 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-controller-manager-operator_cluster-cloud-controller-manager-operator-758cf9d97b-74dgz_af7078c2-ad4f-415b-ba2f-77f5145c3b3f/cluster-cloud-controller-manager/1.log" Dec 05 11:53:34.534180 master-0 kubenswrapper[24928]: I1205 11:53:34.533585 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-controller-manager-operator_cluster-cloud-controller-manager-operator-758cf9d97b-74dgz_af7078c2-ad4f-415b-ba2f-77f5145c3b3f/cluster-cloud-controller-manager/0.log" Dec 05 11:53:34.552580 master-0 kubenswrapper[24928]: I1205 11:53:34.552527 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-controller-manager-operator_cluster-cloud-controller-manager-operator-758cf9d97b-74dgz_af7078c2-ad4f-415b-ba2f-77f5145c3b3f/config-sync-controllers/1.log" Dec 05 11:53:34.553759 master-0 kubenswrapper[24928]: I1205 11:53:34.553195 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-controller-manager-operator_cluster-cloud-controller-manager-operator-758cf9d97b-74dgz_af7078c2-ad4f-415b-ba2f-77f5145c3b3f/config-sync-controllers/0.log" Dec 05 11:53:34.575059 master-0 kubenswrapper[24928]: I1205 11:53:34.574984 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-controller-manager-operator_cluster-cloud-controller-manager-operator-758cf9d97b-74dgz_af7078c2-ad4f-415b-ba2f-77f5145c3b3f/kube-rbac-proxy/0.log" Dec 05 11:53:36.809014 master-0 kubenswrapper[24928]: I1205 11:53:36.808944 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-credential-operator_cloud-credential-operator-698c598cfc-rgc4p_ee0b4a7f-1b96-4304-bef0-fb575e0e431c/kube-rbac-proxy/0.log" Dec 05 11:53:36.852707 master-0 kubenswrapper[24928]: I1205 11:53:36.852622 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cloud-credential-operator_cloud-credential-operator-698c598cfc-rgc4p_ee0b4a7f-1b96-4304-bef0-fb575e0e431c/cloud-credential-operator/0.log" Dec 05 11:53:38.601139 master-0 kubenswrapper[24928]: I1205 11:53:38.593232 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-config-operator_openshift-config-operator-68758cbcdb-dnpcv_c22d947f-a5b6-4f24-b142-dd201c46293b/openshift-config-operator/1.log" Dec 05 11:53:38.601139 master-0 kubenswrapper[24928]: I1205 11:53:38.598778 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-config-operator_openshift-config-operator-68758cbcdb-dnpcv_c22d947f-a5b6-4f24-b142-dd201c46293b/openshift-config-operator/0.log" Dec 05 11:53:38.632572 master-0 kubenswrapper[24928]: I1205 11:53:38.632235 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-config-operator_openshift-config-operator-68758cbcdb-dnpcv_c22d947f-a5b6-4f24-b142-dd201c46293b/openshift-api/0.log" Dec 05 11:53:39.557452 master-0 kubenswrapper[24928]: I1205 11:53:39.557377 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-54dbc87ccb-m7p5f_f3fa06fd-e24e-4bb1-8d20-be16c535d89f/console-operator/0.log" Dec 05 11:53:40.300288 master-0 kubenswrapper[24928]: I1205 11:53:40.300234 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-78d584df9-x54pl_e52dfad5-4a11-479b-8e34-276840f26282/console/0.log" Dec 05 11:53:40.351476 master-0 kubenswrapper[24928]: I1205 11:53:40.351411 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_downloads-69cd4c69bf-d9jtn_3c708092-6503-4a61-9230-617f8a1e2d19/download-server/0.log" Dec 05 11:53:41.116795 master-0 kubenswrapper[24928]: I1205 11:53:41.116106 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_cluster-storage-operator-dcf7fc84b-9rzps_58b95ae4-7f74-4777-8441-0a0ae28199e9/cluster-storage-operator/0.log" Dec 05 11:53:41.132091 master-0 kubenswrapper[24928]: I1205 11:53:41.132031 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/4.log" Dec 05 11:53:41.132797 master-0 kubenswrapper[24928]: I1205 11:53:41.132752 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-6b958b6f94-lgn6v_e27c0798-ec1c-43cd-b81b-f77f2f11ad0f/snapshot-controller/3.log" Dec 05 11:53:41.167279 master-0 kubenswrapper[24928]: I1205 11:53:41.164629 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-storage-operator_csi-snapshot-controller-operator-6bc8656fdc-vd94f_87909f47-f2d7-46f8-a1c8-27336cdcce5d/csi-snapshot-controller-operator/0.log" Dec 05 11:53:41.853326 master-0 kubenswrapper[24928]: I1205 11:53:41.852485 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns-operator_dns-operator-7c56cf9b74-x6t9h_5a1bdc70-6412-47e0-8330-04d796cc8d55/dns-operator/0.log" Dec 05 11:53:41.877472 master-0 kubenswrapper[24928]: I1205 11:53:41.876175 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns-operator_dns-operator-7c56cf9b74-x6t9h_5a1bdc70-6412-47e0-8330-04d796cc8d55/kube-rbac-proxy/0.log" Dec 05 11:53:42.186531 master-0 kubenswrapper[24928]: I1205 11:53:42.186241 24928 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-must-gather-h7tv2/perf-node-gather-daemonset-tz9gc" Dec 05 11:53:42.569392 master-0 kubenswrapper[24928]: I1205 11:53:42.569213 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-4vxng_58baad85-de54-49e7-a13e-d470d9c50d11/dns/0.log" Dec 05 11:53:42.582023 master-0 kubenswrapper[24928]: I1205 11:53:42.581971 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_dns-default-4vxng_58baad85-de54-49e7-a13e-d470d9c50d11/kube-rbac-proxy/0.log" Dec 05 11:53:42.601840 master-0 kubenswrapper[24928]: I1205 11:53:42.601750 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-dns_node-resolver-qkccw_baee05cb-62ad-4eda-8a13-e317eae840e5/dns-node-resolver/0.log" Dec 05 11:53:43.092833 master-0 kubenswrapper[24928]: I1205 11:53:43.092741 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h7tv2/master-0-debug-629dn" event={"ID":"6d44374e-1ad4-47df-a596-e8c2b974783a","Type":"ContainerStarted","Data":"3ce7993480a4af56dd7f4d013e19948184eb5eae21d07f829d689b342a489aa8"} Dec 05 11:53:43.373397 master-0 kubenswrapper[24928]: I1205 11:53:43.373285 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c_f7a08359-0379-4364-8b0c-ddb58ff605f4/etcd-operator/0.log" Dec 05 11:53:43.384779 master-0 kubenswrapper[24928]: I1205 11:53:43.382612 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd-operator_etcd-operator-5bf4d88c6f-n8t5c_f7a08359-0379-4364-8b0c-ddb58ff605f4/etcd-operator/1.log" Dec 05 11:53:44.309661 master-0 kubenswrapper[24928]: I1205 11:53:44.309605 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcdctl/0.log" Dec 05 11:53:45.412154 master-0 kubenswrapper[24928]: I1205 11:53:45.412091 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd/0.log" Dec 05 11:53:45.429020 master-0 kubenswrapper[24928]: I1205 11:53:45.428955 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd-metrics/0.log" Dec 05 11:53:45.439347 master-0 kubenswrapper[24928]: I1205 11:53:45.439279 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd-readyz/0.log" Dec 05 11:53:45.452817 master-0 kubenswrapper[24928]: I1205 11:53:45.452764 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd-rev/0.log" Dec 05 11:53:45.466194 master-0 kubenswrapper[24928]: I1205 11:53:45.466129 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/setup/0.log" Dec 05 11:53:45.489629 master-0 kubenswrapper[24928]: I1205 11:53:45.489490 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd-ensure-env-vars/0.log" Dec 05 11:53:45.646525 master-0 kubenswrapper[24928]: I1205 11:53:45.646470 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_etcd-master-0_58d12e893528ad53a994f10901a644ea/etcd-resources-copy/0.log" Dec 05 11:53:45.699481 master-0 kubenswrapper[24928]: I1205 11:53:45.699282 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-1-master-0_e149cbe7-1ec3-4a06-af61-3a8906b8e9ef/installer/0.log" Dec 05 11:53:45.721412 master-0 kubenswrapper[24928]: I1205 11:53:45.721340 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-2-master-0_68058d25-5888-495f-8571-453f0bf919d9/installer/0.log" Dec 05 11:53:45.763988 master-0 kubenswrapper[24928]: I1205 11:53:45.763801 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-etcd_installer-2-retry-1-master-0_ee0f69ec-1862-4016-bf9d-7edf6554f473/installer/0.log" Dec 05 11:53:46.902535 master-0 kubenswrapper[24928]: I1205 11:53:46.902446 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_cluster-image-registry-operator-6fb9f88b7-f29mb_f7b29f89-e42d-4e53-ad14-05efdce933f0/cluster-image-registry-operator/0.log" Dec 05 11:53:46.921806 master-0 kubenswrapper[24928]: I1205 11:53:46.921688 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-image-registry_node-ca-np6r8_dbafe605-f87b-47c6-9aa4-ae996044461c/node-ca/0.log" Dec 05 11:53:47.580196 master-0 kubenswrapper[24928]: I1205 11:53:47.580118 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/3.log" Dec 05 11:53:47.585636 master-0 kubenswrapper[24928]: I1205 11:53:47.585593 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/ingress-operator/4.log" Dec 05 11:53:47.601403 master-0 kubenswrapper[24928]: I1205 11:53:47.600510 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-operator_ingress-operator-8649c48786-cgt5x_22676fac-b770-4937-9bee-7478bd1babb7/kube-rbac-proxy/0.log" Dec 05 11:53:48.347067 master-0 kubenswrapper[24928]: I1205 11:53:48.346987 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress-canary_ingress-canary-knq92_b7508cd3-7421-4ab9-be1f-318db2853596/serve-healthcheck-canary/0.log" Dec 05 11:53:49.001962 master-0 kubenswrapper[24928]: I1205 11:53:49.001900 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-insights_insights-operator-55965856b6-2sxv7_65e06ca6-1f15-43ec-b8dc-a300e1cf83bd/insights-operator/0.log" Dec 05 11:53:50.896134 master-0 kubenswrapper[24928]: I1205 11:53:50.896086 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_880060c7-79ed-428b-be59-ff663617615b/alertmanager/0.log" Dec 05 11:53:50.924602 master-0 kubenswrapper[24928]: I1205 11:53:50.924520 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_880060c7-79ed-428b-be59-ff663617615b/config-reloader/0.log" Dec 05 11:53:50.957698 master-0 kubenswrapper[24928]: I1205 11:53:50.957652 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_880060c7-79ed-428b-be59-ff663617615b/kube-rbac-proxy-web/0.log" Dec 05 11:53:50.977914 master-0 kubenswrapper[24928]: I1205 11:53:50.977867 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_880060c7-79ed-428b-be59-ff663617615b/kube-rbac-proxy/0.log" Dec 05 11:53:50.992872 master-0 kubenswrapper[24928]: I1205 11:53:50.992813 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_880060c7-79ed-428b-be59-ff663617615b/kube-rbac-proxy-metric/0.log" Dec 05 11:53:51.007511 master-0 kubenswrapper[24928]: I1205 11:53:51.007459 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_880060c7-79ed-428b-be59-ff663617615b/prom-label-proxy/0.log" Dec 05 11:53:51.025232 master-0 kubenswrapper[24928]: I1205 11:53:51.025159 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_alertmanager-main-0_880060c7-79ed-428b-be59-ff663617615b/init-config-reloader/0.log" Dec 05 11:53:51.095236 master-0 kubenswrapper[24928]: I1205 11:53:51.095174 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_cluster-monitoring-operator-7ff994598c-kq8qr_a722cda9-29a0-4b7f-8e1d-9a8950ed765a/cluster-monitoring-operator/0.log" Dec 05 11:53:51.121184 master-0 kubenswrapper[24928]: I1205 11:53:51.121129 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_kube-state-metrics-5857974f64-xj7pj_54f9f9d4-8c31-4636-8010-22ee11b9b323/kube-state-metrics/0.log" Dec 05 11:53:51.141775 master-0 kubenswrapper[24928]: I1205 11:53:51.141710 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_kube-state-metrics-5857974f64-xj7pj_54f9f9d4-8c31-4636-8010-22ee11b9b323/kube-rbac-proxy-main/0.log" Dec 05 11:53:51.159739 master-0 kubenswrapper[24928]: I1205 11:53:51.159193 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_kube-state-metrics-5857974f64-xj7pj_54f9f9d4-8c31-4636-8010-22ee11b9b323/kube-rbac-proxy-self/0.log" Dec 05 11:53:51.175946 master-0 kubenswrapper[24928]: I1205 11:53:51.175897 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_metrics-server-64494f74c5-sqgmf_1b2240bc-13f2-457d-b9b1-72ca9f8035d1/metrics-server/0.log" Dec 05 11:53:51.196102 master-0 kubenswrapper[24928]: I1205 11:53:51.195552 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_monitoring-plugin-54d7d75457-2k7b8_e6acb017-11a4-426a-ae9e-1590f23e8090/monitoring-plugin/0.log" Dec 05 11:53:51.213741 master-0 kubenswrapper[24928]: I1205 11:53:51.213653 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-bmqsb_21a9bd5a-3ade-48ef-8004-e0492daa85ba/node-exporter/0.log" Dec 05 11:53:51.223824 master-0 kubenswrapper[24928]: I1205 11:53:51.223780 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-bmqsb_21a9bd5a-3ade-48ef-8004-e0492daa85ba/kube-rbac-proxy/0.log" Dec 05 11:53:51.243220 master-0 kubenswrapper[24928]: I1205 11:53:51.243172 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_node-exporter-bmqsb_21a9bd5a-3ade-48ef-8004-e0492daa85ba/init-textfile/0.log" Dec 05 11:53:51.262787 master-0 kubenswrapper[24928]: I1205 11:53:51.262738 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-5974b6b869-9p5mt_d9ff1ae2-8547-42db-9fd0-f4782589ca18/kube-rbac-proxy-main/0.log" Dec 05 11:53:51.277458 master-0 kubenswrapper[24928]: I1205 11:53:51.277360 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-5974b6b869-9p5mt_d9ff1ae2-8547-42db-9fd0-f4782589ca18/kube-rbac-proxy-self/0.log" Dec 05 11:53:51.301136 master-0 kubenswrapper[24928]: I1205 11:53:51.301067 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_openshift-state-metrics-5974b6b869-9p5mt_d9ff1ae2-8547-42db-9fd0-f4782589ca18/openshift-state-metrics/0.log" Dec 05 11:53:51.347456 master-0 kubenswrapper[24928]: I1205 11:53:51.347387 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_97818d19-9a0b-40a3-a075-a3ed60eebe62/prometheus/0.log" Dec 05 11:53:51.363283 master-0 kubenswrapper[24928]: I1205 11:53:51.363227 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_97818d19-9a0b-40a3-a075-a3ed60eebe62/config-reloader/0.log" Dec 05 11:53:51.382243 master-0 kubenswrapper[24928]: I1205 11:53:51.382195 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_97818d19-9a0b-40a3-a075-a3ed60eebe62/thanos-sidecar/0.log" Dec 05 11:53:51.400774 master-0 kubenswrapper[24928]: I1205 11:53:51.400571 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_97818d19-9a0b-40a3-a075-a3ed60eebe62/kube-rbac-proxy-web/0.log" Dec 05 11:53:51.422964 master-0 kubenswrapper[24928]: I1205 11:53:51.422856 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_97818d19-9a0b-40a3-a075-a3ed60eebe62/kube-rbac-proxy/0.log" Dec 05 11:53:51.442342 master-0 kubenswrapper[24928]: I1205 11:53:51.442220 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_97818d19-9a0b-40a3-a075-a3ed60eebe62/kube-rbac-proxy-thanos/0.log" Dec 05 11:53:51.463849 master-0 kubenswrapper[24928]: I1205 11:53:51.463797 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-k8s-0_97818d19-9a0b-40a3-a075-a3ed60eebe62/init-config-reloader/0.log" Dec 05 11:53:51.488945 master-0 kubenswrapper[24928]: I1205 11:53:51.488898 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-6c74d9cb9f-r787z_b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e/prometheus-operator/0.log" Dec 05 11:53:51.516725 master-0 kubenswrapper[24928]: I1205 11:53:51.516654 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-6c74d9cb9f-r787z_b02b2ea9-2bd2-4ea8-a96b-16d5f13cf19e/kube-rbac-proxy/0.log" Dec 05 11:53:51.539732 master-0 kubenswrapper[24928]: I1205 11:53:51.539692 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_prometheus-operator-admission-webhook-7c85c4dffd-vjvbz_0babac16-164a-405a-a86c-30524118f046/prometheus-operator-admission-webhook/0.log" Dec 05 11:53:51.573485 master-0 kubenswrapper[24928]: I1205 11:53:51.573444 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-86cb595668-52qnw_384f1c47-f511-4b58-aa7f-71aef6ef91a9/telemeter-client/0.log" Dec 05 11:53:51.590771 master-0 kubenswrapper[24928]: I1205 11:53:51.590712 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-86cb595668-52qnw_384f1c47-f511-4b58-aa7f-71aef6ef91a9/reload/0.log" Dec 05 11:53:51.605288 master-0 kubenswrapper[24928]: I1205 11:53:51.605090 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_telemeter-client-86cb595668-52qnw_384f1c47-f511-4b58-aa7f-71aef6ef91a9/kube-rbac-proxy/0.log" Dec 05 11:53:51.632938 master-0 kubenswrapper[24928]: I1205 11:53:51.632885 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-598896584f-9pd95_c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7/thanos-query/0.log" Dec 05 11:53:51.648908 master-0 kubenswrapper[24928]: I1205 11:53:51.648869 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-598896584f-9pd95_c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7/kube-rbac-proxy-web/0.log" Dec 05 11:53:51.670297 master-0 kubenswrapper[24928]: I1205 11:53:51.670237 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-598896584f-9pd95_c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7/kube-rbac-proxy/0.log" Dec 05 11:53:51.687365 master-0 kubenswrapper[24928]: I1205 11:53:51.687231 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-598896584f-9pd95_c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7/prom-label-proxy/0.log" Dec 05 11:53:51.706911 master-0 kubenswrapper[24928]: I1205 11:53:51.706868 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-598896584f-9pd95_c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7/kube-rbac-proxy-rules/0.log" Dec 05 11:53:51.737897 master-0 kubenswrapper[24928]: I1205 11:53:51.737786 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-monitoring_thanos-querier-598896584f-9pd95_c4b24b10-85c9-4ef1-8cd3-3afc0cc65af7/kube-rbac-proxy-metrics/0.log" Dec 05 11:53:54.051216 master-0 kubenswrapper[24928]: I1205 11:53:54.051165 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-fpl59_2a019084-2c01-4311-81d4-df4e5fe47107/controller/0.log" Dec 05 11:53:54.066486 master-0 kubenswrapper[24928]: I1205 11:53:54.066342 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-fpl59_2a019084-2c01-4311-81d4-df4e5fe47107/kube-rbac-proxy/0.log" Dec 05 11:53:54.117481 master-0 kubenswrapper[24928]: I1205 11:53:54.117408 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/controller/0.log" Dec 05 11:53:55.487233 master-0 kubenswrapper[24928]: I1205 11:53:55.487170 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/frr/0.log" Dec 05 11:53:55.652629 master-0 kubenswrapper[24928]: I1205 11:53:55.652580 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/reloader/0.log" Dec 05 11:53:55.661927 master-0 kubenswrapper[24928]: I1205 11:53:55.661883 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/frr-metrics/0.log" Dec 05 11:53:55.675695 master-0 kubenswrapper[24928]: I1205 11:53:55.675647 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/kube-rbac-proxy/0.log" Dec 05 11:53:55.685743 master-0 kubenswrapper[24928]: I1205 11:53:55.685704 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/kube-rbac-proxy-frr/0.log" Dec 05 11:53:55.698403 master-0 kubenswrapper[24928]: I1205 11:53:55.698315 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/cp-frr-files/0.log" Dec 05 11:53:55.715201 master-0 kubenswrapper[24928]: I1205 11:53:55.715080 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/cp-reloader/0.log" Dec 05 11:53:55.733724 master-0 kubenswrapper[24928]: I1205 11:53:55.733669 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-2cn6b_fec8fc95-cfe1-4f43-b693-7a11680da86a/cp-metrics/0.log" Dec 05 11:53:55.750830 master-0 kubenswrapper[24928]: I1205 11:53:55.750690 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-dlsnb_a65480f6-aa9d-45b5-b3ba-541e87b7b145/frr-k8s-webhook-server/0.log" Dec 05 11:53:55.781625 master-0 kubenswrapper[24928]: I1205 11:53:55.781561 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9d5bd9bc7-q878m_697f986f-bfa7-4e85-9d4b-5f07a4675228/manager/0.log" Dec 05 11:53:55.801457 master-0 kubenswrapper[24928]: I1205 11:53:55.801368 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5f77dd7bb4-xmg4x_c39fe00d-1550-4d88-8b8a-95439d6e58e1/webhook-server/0.log" Dec 05 11:53:56.198323 master-0 kubenswrapper[24928]: I1205 11:53:56.198247 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9stls_b953bb78-052f-4b26-94d3-da9bab7a9fdc/speaker/0.log" Dec 05 11:53:56.216509 master-0 kubenswrapper[24928]: I1205 11:53:56.216408 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-9stls_b953bb78-052f-4b26-94d3-da9bab7a9fdc/kube-rbac-proxy/0.log" Dec 05 11:53:58.822669 master-0 kubenswrapper[24928]: I1205 11:53:58.820957 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-node-tuning-operator_cluster-node-tuning-operator-85cff47f46-qwx2p_eb290494-a456-4f0e-9afc-f20abab1a1bf/cluster-node-tuning-operator/0.log" Dec 05 11:53:58.845885 master-0 kubenswrapper[24928]: I1205 11:53:58.845816 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-node-tuning-operator_tuned-hvh88_33616deb-ea10-4a38-8681-ab023b526b11/tuned/0.log" Dec 05 11:54:00.557560 master-0 kubenswrapper[24928]: I1205 11:54:00.556446 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl_444f8808-e454-4015-9e20-429e715a08c7/kube-apiserver-operator/0.log" Dec 05 11:54:00.557560 master-0 kubenswrapper[24928]: I1205 11:54:00.556550 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver-operator_kube-apiserver-operator-765d9ff747-p57fl_444f8808-e454-4015-9e20-429e715a08c7/kube-apiserver-operator/1.log" Dec 05 11:54:01.286468 master-0 kubenswrapper[24928]: I1205 11:54:01.286390 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-1-master-0_5d8c854f-31f8-42d3-b28c-82c7d1b14ee9/installer/0.log" Dec 05 11:54:01.313532 master-0 kubenswrapper[24928]: I1205 11:54:01.313001 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-3-master-0_75ba20a1-b322-4ce8-88d2-d8ee210cc27a/installer/0.log" Dec 05 11:54:01.339659 master-0 kubenswrapper[24928]: I1205 11:54:01.339480 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_installer-5-master-0_4d237cdd-9aa8-4a12-b244-1dc1b4c456d0/installer/0.log" Dec 05 11:54:02.190194 master-0 kubenswrapper[24928]: I1205 11:54:02.190099 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-master-0_a369cadf0161d66f2936cdea3ded59b7/kube-apiserver/0.log" Dec 05 11:54:02.204175 master-0 kubenswrapper[24928]: I1205 11:54:02.202439 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-master-0_a369cadf0161d66f2936cdea3ded59b7/kube-apiserver-cert-syncer/0.log" Dec 05 11:54:02.217829 master-0 kubenswrapper[24928]: I1205 11:54:02.217758 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-master-0_a369cadf0161d66f2936cdea3ded59b7/kube-apiserver-cert-regeneration-controller/0.log" Dec 05 11:54:02.239334 master-0 kubenswrapper[24928]: I1205 11:54:02.239282 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-master-0_a369cadf0161d66f2936cdea3ded59b7/kube-apiserver-insecure-readyz/0.log" Dec 05 11:54:02.258210 master-0 kubenswrapper[24928]: I1205 11:54:02.258125 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-master-0_a369cadf0161d66f2936cdea3ded59b7/kube-apiserver-check-endpoints/0.log" Dec 05 11:54:02.273613 master-0 kubenswrapper[24928]: I1205 11:54:02.273513 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-master-0_a369cadf0161d66f2936cdea3ded59b7/setup/0.log" Dec 05 11:54:05.567330 master-0 kubenswrapper[24928]: I1205 11:54:05.567281 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-catalogd_catalogd-controller-manager-7cc89f4c4c-lth87_a4fd453c-e667-4bdc-aa9e-3d95ff707200/kube-rbac-proxy/0.log" Dec 05 11:54:05.604518 master-0 kubenswrapper[24928]: I1205 11:54:05.604466 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-catalogd_catalogd-controller-manager-7cc89f4c4c-lth87_a4fd453c-e667-4bdc-aa9e-3d95ff707200/manager/2.log" Dec 05 11:54:05.663216 master-0 kubenswrapper[24928]: I1205 11:54:05.663141 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-catalogd_catalogd-controller-manager-7cc89f4c4c-lth87_a4fd453c-e667-4bdc-aa9e-3d95ff707200/manager/1.log" Dec 05 11:54:06.526316 master-0 kubenswrapper[24928]: I1205 11:54:06.526206 24928 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h7tv2/master-0-debug-629dn" podStartSLOduration=24.740728286 podStartE2EDuration="40.526188014s" podCreationTimestamp="2025-12-05 11:53:26 +0000 UTC" firstStartedPulling="2025-12-05 11:53:26.70455572 +0000 UTC m=+3966.707749571" lastFinishedPulling="2025-12-05 11:53:42.490015438 +0000 UTC m=+3982.493209299" observedRunningTime="2025-12-05 11:53:43.124512255 +0000 UTC m=+3983.127706126" watchObservedRunningTime="2025-12-05 11:54:06.526188014 +0000 UTC m=+4006.529381855" Dec 05 11:54:06.528974 master-0 kubenswrapper[24928]: I1205 11:54:06.528906 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b92xr"] Dec 05 11:54:06.531221 master-0 kubenswrapper[24928]: I1205 11:54:06.531198 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b92xr" Dec 05 11:54:06.691397 master-0 kubenswrapper[24928]: I1205 11:54:06.691317 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb832396-def1-4c21-b7ab-d6da7af574bd-catalog-content\") pod \"redhat-operators-b92xr\" (UID: \"fb832396-def1-4c21-b7ab-d6da7af574bd\") " pod="openshift-marketplace/redhat-operators-b92xr" Dec 05 11:54:06.692154 master-0 kubenswrapper[24928]: I1205 11:54:06.691451 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb832396-def1-4c21-b7ab-d6da7af574bd-utilities\") pod \"redhat-operators-b92xr\" (UID: \"fb832396-def1-4c21-b7ab-d6da7af574bd\") " pod="openshift-marketplace/redhat-operators-b92xr" Dec 05 11:54:06.692154 master-0 kubenswrapper[24928]: I1205 11:54:06.691730 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svwcz\" (UniqueName: \"kubernetes.io/projected/fb832396-def1-4c21-b7ab-d6da7af574bd-kube-api-access-svwcz\") pod \"redhat-operators-b92xr\" (UID: \"fb832396-def1-4c21-b7ab-d6da7af574bd\") " pod="openshift-marketplace/redhat-operators-b92xr" Dec 05 11:54:06.793599 master-0 kubenswrapper[24928]: I1205 11:54:06.793505 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb832396-def1-4c21-b7ab-d6da7af574bd-utilities\") pod \"redhat-operators-b92xr\" (UID: \"fb832396-def1-4c21-b7ab-d6da7af574bd\") " pod="openshift-marketplace/redhat-operators-b92xr" Dec 05 11:54:06.793818 master-0 kubenswrapper[24928]: I1205 11:54:06.793790 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svwcz\" (UniqueName: \"kubernetes.io/projected/fb832396-def1-4c21-b7ab-d6da7af574bd-kube-api-access-svwcz\") pod \"redhat-operators-b92xr\" (UID: \"fb832396-def1-4c21-b7ab-d6da7af574bd\") " pod="openshift-marketplace/redhat-operators-b92xr" Dec 05 11:54:06.794162 master-0 kubenswrapper[24928]: I1205 11:54:06.793932 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb832396-def1-4c21-b7ab-d6da7af574bd-catalog-content\") pod \"redhat-operators-b92xr\" (UID: \"fb832396-def1-4c21-b7ab-d6da7af574bd\") " pod="openshift-marketplace/redhat-operators-b92xr" Dec 05 11:54:06.794677 master-0 kubenswrapper[24928]: I1205 11:54:06.794633 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb832396-def1-4c21-b7ab-d6da7af574bd-catalog-content\") pod \"redhat-operators-b92xr\" (UID: \"fb832396-def1-4c21-b7ab-d6da7af574bd\") " pod="openshift-marketplace/redhat-operators-b92xr" Dec 05 11:54:06.795016 master-0 kubenswrapper[24928]: I1205 11:54:06.794987 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb832396-def1-4c21-b7ab-d6da7af574bd-utilities\") pod \"redhat-operators-b92xr\" (UID: \"fb832396-def1-4c21-b7ab-d6da7af574bd\") " pod="openshift-marketplace/redhat-operators-b92xr" Dec 05 11:54:07.198604 master-0 kubenswrapper[24928]: I1205 11:54:07.185123 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svwcz\" (UniqueName: \"kubernetes.io/projected/fb832396-def1-4c21-b7ab-d6da7af574bd-kube-api-access-svwcz\") pod \"redhat-operators-b92xr\" (UID: \"fb832396-def1-4c21-b7ab-d6da7af574bd\") " pod="openshift-marketplace/redhat-operators-b92xr" Dec 05 11:54:07.252074 master-0 kubenswrapper[24928]: I1205 11:54:07.251999 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b92xr"] Dec 05 11:54:07.452688 master-0 kubenswrapper[24928]: I1205 11:54:07.452446 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b92xr" Dec 05 11:54:07.909001 master-0 kubenswrapper[24928]: I1205 11:54:07.908943 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-4l8x5_f22dc7c0-3179-4346-8a71-865c2a0bb10e/cert-manager-controller/0.log" Dec 05 11:54:07.932865 master-0 kubenswrapper[24928]: I1205 11:54:07.931971 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-jkch2_dc0f03cc-abed-444e-8039-d6a9ed2c936b/cert-manager-cainjector/0.log" Dec 05 11:54:07.973225 master-0 kubenswrapper[24928]: I1205 11:54:07.973158 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-29nx4_5c4c235c-46d5-43e6-a12b-e93dd86a2fcf/cert-manager-webhook/0.log" Dec 05 11:54:08.100141 master-0 kubenswrapper[24928]: I1205 11:54:08.099922 24928 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b92xr"] Dec 05 11:54:08.390303 master-0 kubenswrapper[24928]: I1205 11:54:08.390187 24928 generic.go:334] "Generic (PLEG): container finished" podID="6d44374e-1ad4-47df-a596-e8c2b974783a" containerID="3ce7993480a4af56dd7f4d013e19948184eb5eae21d07f829d689b342a489aa8" exitCode=0 Dec 05 11:54:08.390303 master-0 kubenswrapper[24928]: I1205 11:54:08.390248 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h7tv2/master-0-debug-629dn" event={"ID":"6d44374e-1ad4-47df-a596-e8c2b974783a","Type":"ContainerDied","Data":"3ce7993480a4af56dd7f4d013e19948184eb5eae21d07f829d689b342a489aa8"} Dec 05 11:54:09.091262 master-0 kubenswrapper[24928]: I1205 11:54:09.091182 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7_0fb13d51-751c-4313-99f3-27db02ea0911/extract/0.log" Dec 05 11:54:09.107505 master-0 kubenswrapper[24928]: I1205 11:54:09.107457 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7_0fb13d51-751c-4313-99f3-27db02ea0911/util/0.log" Dec 05 11:54:09.127850 master-0 kubenswrapper[24928]: I1205 11:54:09.127773 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eaffmmf7_0fb13d51-751c-4313-99f3-27db02ea0911/pull/0.log" Dec 05 11:54:09.149892 master-0 kubenswrapper[24928]: I1205 11:54:09.149826 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5cd89994b5-ssmd2_5017646a-fb39-440c-afe5-a4d6cee68ed4/manager/0.log" Dec 05 11:54:09.165073 master-0 kubenswrapper[24928]: I1205 11:54:09.164843 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-5cd89994b5-ssmd2_5017646a-fb39-440c-afe5-a4d6cee68ed4/kube-rbac-proxy/0.log" Dec 05 11:54:09.215372 master-0 kubenswrapper[24928]: I1205 11:54:09.215291 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-f8856dd79-7582v_4f712f48-af26-4222-a231-ac989f78d18f/manager/0.log" Dec 05 11:54:09.225375 master-0 kubenswrapper[24928]: I1205 11:54:09.225191 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-f8856dd79-7582v_4f712f48-af26-4222-a231-ac989f78d18f/kube-rbac-proxy/0.log" Dec 05 11:54:09.248180 master-0 kubenswrapper[24928]: I1205 11:54:09.248129 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84bc9f68f5-t8l7w_44343d25-46da-40a9-b0be-94861421d403/manager/0.log" Dec 05 11:54:09.258780 master-0 kubenswrapper[24928]: I1205 11:54:09.258727 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84bc9f68f5-t8l7w_44343d25-46da-40a9-b0be-94861421d403/kube-rbac-proxy/0.log" Dec 05 11:54:09.378582 master-0 kubenswrapper[24928]: I1205 11:54:09.378452 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78cd4f7769-xpcsc_216bdcf3-b763-4293-8ac4-3e7eeeae0b5d/manager/0.log" Dec 05 11:54:09.395177 master-0 kubenswrapper[24928]: I1205 11:54:09.395123 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-78cd4f7769-xpcsc_216bdcf3-b763-4293-8ac4-3e7eeeae0b5d/kube-rbac-proxy/0.log" Dec 05 11:54:09.415069 master-0 kubenswrapper[24928]: I1205 11:54:09.415025 24928 generic.go:334] "Generic (PLEG): container finished" podID="fb832396-def1-4c21-b7ab-d6da7af574bd" containerID="dc59a981327204fbbfdb3bb43188b446bc1967c9b9928645ee83299152d2d049" exitCode=0 Dec 05 11:54:09.415604 master-0 kubenswrapper[24928]: I1205 11:54:09.415517 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b92xr" event={"ID":"fb832396-def1-4c21-b7ab-d6da7af574bd","Type":"ContainerDied","Data":"dc59a981327204fbbfdb3bb43188b446bc1967c9b9928645ee83299152d2d049"} Dec 05 11:54:09.415685 master-0 kubenswrapper[24928]: I1205 11:54:09.415648 24928 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b92xr" event={"ID":"fb832396-def1-4c21-b7ab-d6da7af574bd","Type":"ContainerStarted","Data":"17af618b59e071a3dc21e364ac5b26043e535cd98f601bb1c826b78068e559c4"} Dec 05 11:54:09.425120 master-0 kubenswrapper[24928]: I1205 11:54:09.425075 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-7fd96594c7-5k6gc_1d175522-0c08-40c5-a309-f1820406ea2f/manager/0.log" Dec 05 11:54:09.446580 master-0 kubenswrapper[24928]: I1205 11:54:09.446533 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-7fd96594c7-5k6gc_1d175522-0c08-40c5-a309-f1820406ea2f/kube-rbac-proxy/0.log" Dec 05 11:54:09.468674 master-0 kubenswrapper[24928]: I1205 11:54:09.468629 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-f6cc97788-5lr6c_be90920f-159a-4fac-9746-70846d306f22/manager/0.log" Dec 05 11:54:09.486891 master-0 kubenswrapper[24928]: I1205 11:54:09.486857 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-f6cc97788-5lr6c_be90920f-159a-4fac-9746-70846d306f22/kube-rbac-proxy/0.log" Dec 05 11:54:09.526189 master-0 kubenswrapper[24928]: I1205 11:54:09.526145 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/master-0-debug-629dn" Dec 05 11:54:09.572116 master-0 kubenswrapper[24928]: I1205 11:54:09.572058 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mntvb\" (UniqueName: \"kubernetes.io/projected/6d44374e-1ad4-47df-a596-e8c2b974783a-kube-api-access-mntvb\") pod \"6d44374e-1ad4-47df-a596-e8c2b974783a\" (UID: \"6d44374e-1ad4-47df-a596-e8c2b974783a\") " Dec 05 11:54:09.573003 master-0 kubenswrapper[24928]: I1205 11:54:09.572956 24928 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d44374e-1ad4-47df-a596-e8c2b974783a-host\") pod \"6d44374e-1ad4-47df-a596-e8c2b974783a\" (UID: \"6d44374e-1ad4-47df-a596-e8c2b974783a\") " Dec 05 11:54:09.573270 master-0 kubenswrapper[24928]: I1205 11:54:09.573221 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d44374e-1ad4-47df-a596-e8c2b974783a-host" (OuterVolumeSpecName: "host") pod "6d44374e-1ad4-47df-a596-e8c2b974783a" (UID: "6d44374e-1ad4-47df-a596-e8c2b974783a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 11:54:09.574353 master-0 kubenswrapper[24928]: I1205 11:54:09.574326 24928 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6d44374e-1ad4-47df-a596-e8c2b974783a-host\") on node \"master-0\" DevicePath \"\"" Dec 05 11:54:09.576059 master-0 kubenswrapper[24928]: I1205 11:54:09.576011 24928 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h7tv2/master-0-debug-629dn"] Dec 05 11:54:09.586411 master-0 kubenswrapper[24928]: I1205 11:54:09.586338 24928 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d44374e-1ad4-47df-a596-e8c2b974783a-kube-api-access-mntvb" (OuterVolumeSpecName: "kube-api-access-mntvb") pod "6d44374e-1ad4-47df-a596-e8c2b974783a" (UID: "6d44374e-1ad4-47df-a596-e8c2b974783a"). InnerVolumeSpecName "kube-api-access-mntvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 11:54:09.587415 master-0 kubenswrapper[24928]: I1205 11:54:09.587238 24928 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h7tv2/master-0-debug-629dn"] Dec 05 11:54:09.636175 master-0 kubenswrapper[24928]: I1205 11:54:09.636062 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d9c9d7fd8-4ht2g_da186c71-5e15-45a3-ad76-e51aad43439f/manager/0.log" Dec 05 11:54:09.649837 master-0 kubenswrapper[24928]: I1205 11:54:09.646937 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7d9c9d7fd8-4ht2g_da186c71-5e15-45a3-ad76-e51aad43439f/kube-rbac-proxy/0.log" Dec 05 11:54:09.678452 master-0 kubenswrapper[24928]: I1205 11:54:09.676245 24928 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mntvb\" (UniqueName: \"kubernetes.io/projected/6d44374e-1ad4-47df-a596-e8c2b974783a-kube-api-access-mntvb\") on node \"master-0\" DevicePath \"\"" Dec 05 11:54:09.731541 master-0 kubenswrapper[24928]: I1205 11:54:09.731482 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7c9bfd6967-bhx8z_3d85e846-d44c-49aa-8128-02566193d999/manager/0.log" Dec 05 11:54:09.743836 master-0 kubenswrapper[24928]: I1205 11:54:09.743790 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-7c9bfd6967-bhx8z_3d85e846-d44c-49aa-8128-02566193d999/kube-rbac-proxy/0.log" Dec 05 11:54:09.832957 master-0 kubenswrapper[24928]: I1205 11:54:09.832892 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-58b8dcc5fb-vv6s4_ada63c1e-313b-465a-a556-32b452de65ab/manager/0.log" Dec 05 11:54:09.849180 master-0 kubenswrapper[24928]: I1205 11:54:09.849128 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-58b8dcc5fb-vv6s4_ada63c1e-313b-465a-a556-32b452de65ab/kube-rbac-proxy/0.log" Dec 05 11:54:09.870874 master-0 kubenswrapper[24928]: I1205 11:54:09.870821 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-56f9fbf74b-pwlgc_61b1765d-2b89-4d57-85c0-a7f9afe632fe/manager/0.log" Dec 05 11:54:09.899668 master-0 kubenswrapper[24928]: I1205 11:54:09.899072 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-56f9fbf74b-pwlgc_61b1765d-2b89-4d57-85c0-a7f9afe632fe/kube-rbac-proxy/0.log" Dec 05 11:54:09.933188 master-0 kubenswrapper[24928]: I1205 11:54:09.933124 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-647d75769b-dft2w_442a2392-c758-4713-baa3-3e119ade664c/manager/0.log" Dec 05 11:54:09.946000 master-0 kubenswrapper[24928]: I1205 11:54:09.945619 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-647d75769b-dft2w_442a2392-c758-4713-baa3-3e119ade664c/kube-rbac-proxy/0.log" Dec 05 11:54:10.019799 master-0 kubenswrapper[24928]: I1205 11:54:10.019699 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7cdd6b54fb-9wfjb_5af60df4-8ca0-4bfb-8302-d3f9dcd4ef7a/manager/0.log" Dec 05 11:54:10.035029 master-0 kubenswrapper[24928]: I1205 11:54:10.034849 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7cdd6b54fb-9wfjb_5af60df4-8ca0-4bfb-8302-d3f9dcd4ef7a/kube-rbac-proxy/0.log" Dec 05 11:54:10.122025 master-0 kubenswrapper[24928]: I1205 11:54:10.121959 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-865fc86d5b-z8jv6_fa9681c6-d40e-4075-9230-1ed7431a06fb/manager/0.log" Dec 05 11:54:10.133916 master-0 kubenswrapper[24928]: I1205 11:54:10.133867 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-865fc86d5b-z8jv6_fa9681c6-d40e-4075-9230-1ed7431a06fb/kube-rbac-proxy/0.log" Dec 05 11:54:10.155894 master-0 kubenswrapper[24928]: I1205 11:54:10.155697 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-845b79dc4f-dc9ls_da92e42f-68d5-4f34-bcf5-c5a8e765e383/manager/0.log" Dec 05 11:54:10.176789 master-0 kubenswrapper[24928]: I1205 11:54:10.176744 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-845b79dc4f-dc9ls_da92e42f-68d5-4f34-bcf5-c5a8e765e383/kube-rbac-proxy/0.log" Dec 05 11:54:10.199242 master-0 kubenswrapper[24928]: I1205 11:54:10.197099 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f998f574688x6w_5b821d96-dc12-4794-a863-0cc47d4d8d65/manager/0.log" Dec 05 11:54:10.230524 master-0 kubenswrapper[24928]: I1205 11:54:10.229725 24928 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d44374e-1ad4-47df-a596-e8c2b974783a" path="/var/lib/kubelet/pods/6d44374e-1ad4-47df-a596-e8c2b974783a/volumes" Dec 05 11:54:10.232912 master-0 kubenswrapper[24928]: I1205 11:54:10.232482 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6f998f574688x6w_5b821d96-dc12-4794-a863-0cc47d4d8d65/kube-rbac-proxy/0.log" Dec 05 11:54:10.431865 master-0 kubenswrapper[24928]: I1205 11:54:10.431754 24928 scope.go:117] "RemoveContainer" containerID="3ce7993480a4af56dd7f4d013e19948184eb5eae21d07f829d689b342a489aa8" Dec 05 11:54:10.431865 master-0 kubenswrapper[24928]: I1205 11:54:10.431787 24928 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/master-0-debug-629dn" Dec 05 11:54:10.837468 master-0 kubenswrapper[24928]: I1205 11:54:10.837035 24928 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h7tv2/master-0-debug-vxsl5"] Dec 05 11:54:10.837943 master-0 kubenswrapper[24928]: E1205 11:54:10.837743 24928 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d44374e-1ad4-47df-a596-e8c2b974783a" containerName="container-00" Dec 05 11:54:10.837943 master-0 kubenswrapper[24928]: I1205 11:54:10.837763 24928 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d44374e-1ad4-47df-a596-e8c2b974783a" containerName="container-00" Dec 05 11:54:10.838206 master-0 kubenswrapper[24928]: I1205 11:54:10.838101 24928 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d44374e-1ad4-47df-a596-e8c2b974783a" containerName="container-00" Dec 05 11:54:10.839461 master-0 kubenswrapper[24928]: I1205 11:54:10.839440 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/master-0-debug-vxsl5" Dec 05 11:54:10.900339 master-0 kubenswrapper[24928]: I1205 11:54:10.897875 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kptvc\" (UniqueName: \"kubernetes.io/projected/7b5633af-e935-4bdb-be07-705c8b3539e8-kube-api-access-kptvc\") pod \"master-0-debug-vxsl5\" (UID: \"7b5633af-e935-4bdb-be07-705c8b3539e8\") " pod="openshift-must-gather-h7tv2/master-0-debug-vxsl5" Dec 05 11:54:10.900339 master-0 kubenswrapper[24928]: I1205 11:54:10.898176 24928 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b5633af-e935-4bdb-be07-705c8b3539e8-host\") pod \"master-0-debug-vxsl5\" (UID: \"7b5633af-e935-4bdb-be07-705c8b3539e8\") " pod="openshift-must-gather-h7tv2/master-0-debug-vxsl5" Dec 05 11:54:10.918231 master-0 kubenswrapper[24928]: I1205 11:54:10.918175 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-599cfccd85-8d692_b5f9281b-074b-46fa-8a52-dcdd102f5b8f/manager/0.log" Dec 05 11:54:10.999942 master-0 kubenswrapper[24928]: I1205 11:54:10.999871 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kptvc\" (UniqueName: \"kubernetes.io/projected/7b5633af-e935-4bdb-be07-705c8b3539e8-kube-api-access-kptvc\") pod \"master-0-debug-vxsl5\" (UID: \"7b5633af-e935-4bdb-be07-705c8b3539e8\") " pod="openshift-must-gather-h7tv2/master-0-debug-vxsl5" Dec 05 11:54:11.000213 master-0 kubenswrapper[24928]: I1205 11:54:10.999990 24928 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b5633af-e935-4bdb-be07-705c8b3539e8-host\") pod \"master-0-debug-vxsl5\" (UID: \"7b5633af-e935-4bdb-be07-705c8b3539e8\") " pod="openshift-must-gather-h7tv2/master-0-debug-vxsl5" Dec 05 11:54:11.000213 master-0 kubenswrapper[24928]: I1205 11:54:11.000150 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7b5633af-e935-4bdb-be07-705c8b3539e8-host\") pod \"master-0-debug-vxsl5\" (UID: \"7b5633af-e935-4bdb-be07-705c8b3539e8\") " pod="openshift-must-gather-h7tv2/master-0-debug-vxsl5" Dec 05 11:54:11.034890 master-0 kubenswrapper[24928]: I1205 11:54:11.024645 24928 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kptvc\" (UniqueName: \"kubernetes.io/projected/7b5633af-e935-4bdb-be07-705c8b3539e8-kube-api-access-kptvc\") pod \"master-0-debug-vxsl5\" (UID: \"7b5633af-e935-4bdb-be07-705c8b3539e8\") " pod="openshift-must-gather-h7tv2/master-0-debug-vxsl5" Dec 05 11:54:11.072170 master-0 kubenswrapper[24928]: I1205 11:54:11.068142 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-589d7b4556-v294s_6876062e-bc77-4979-8057-f6b5f389fdcd/operator/0.log" Dec 05 11:54:11.141563 master-0 kubenswrapper[24928]: I1205 11:54:11.141321 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-k64dw_6b418fb9-a9c4-46b7-9be4-0dc52e339e87/registry-server/0.log" Dec 05 11:54:11.183294 master-0 kubenswrapper[24928]: I1205 11:54:11.183167 24928 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h7tv2/master-0-debug-vxsl5" Dec 05 11:54:11.227866 master-0 kubenswrapper[24928]: W1205 11:54:11.227808 24928 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b5633af_e935_4bdb_be07_705c8b3539e8.slice/crio-92e43db331bc8752ee9fc09b4536db2a0354b2700acda4027c44e5d8cd8b8457 WatchSource:0}: Error finding container 92e43db331bc8752ee9fc09b4536db2a0354b2700acda4027c44e5d8cd8b8457: Status 404 returned error can't find the container with id 92e43db331bc8752ee9fc09b4536db2a0354b2700acda4027c44e5d8cd8b8457 Dec 05 11:54:11.245030 master-0 kubenswrapper[24928]: I1205 11:54:11.244991 24928 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-647f96877-gcg9w_68a107e3-b0eb-43fd-844f-817cf01ef767/manager/0.log"